Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2007 Lecture 14b 24 August 2007.

Slides:



Advertisements
Similar presentations
The Structure of Sentences Asian 401
Advertisements

Fall 2004 Lecture Notes #7 EECS 595 / LING 541 / SI 661 Natural Language Processing.
Color Coding Paragraphs for my Lawyers SEEING THE BLUEPRINT of Writing/Arguing (take FN/Cornell notes) By: Mi Linda Valverde Stacey Middle School
Machine Translation Introduction to MT. Dan Jurafsky Machine Translation Fully automatic Helping human translators Enter Source Text: Translation from.
Introduction to phrases & clauses
Introduction to Semantics and Pragmatics. LING NLP 2 NLP tends to focus on: Syntax – Grammars, parsers, parse trees, dependency structures.
1 Words and the Lexicon September 10th 2009 Lecture #3.
ZERO PRONOUN RESOLUTION IN JAPANESE Jeffrey Shu Ling 575 Discourse and Dialogue.
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 23 Jim Martin.
Inducing Information Extraction Systems for New Languages via Cross-Language Projection Ellen Riloff University of Utah Charles Schafer, David Yarowksy.
CMSC 723 / LING 645: Intro to Computational Linguistics September 8, 2004: Dorr MT (continued), MT Evaluation Prof. Bonnie J. Dorr Dr. Christof Monz TA:
Language, Mind, and Brain by Ewa Dabrowska Chapter 2: Language processing: speed and flexibility.
1 Introduction to Computational Linguistics Eleni Miltsakaki AUTH Fall 2005-Lecture 2.
Machine Translation Prof. Alexandros Potamianos Dept. of Electrical & Computer Engineering Technical University of Crete, Greece May 2003.
Resources Primary resources – Lexicons, structured vocabularies – Grammars (in widest sense) – Corpora – Treebanks Secondary resources – Designed for a.
Course Summary LING 575 Fei Xia 03/06/07. Outline Introduction to MT: 1 Major approaches –SMT: 3 –Transfer-based MT: 2 –Hybrid systems: 2 Other topics.
Translation Divergence LING 580MT Fei Xia 1/10/06.
Semi-Automatic Learning of Transfer Rules for Machine Translation of Low-Density Languages Katharina Probst April 5, 2002.
LIN6932 Topics in Computational Linguistics
1 Statistical NLP: Lecture 13 Statistical Alignment and Machine Translation.
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
Computing Science, University of Aberdeen1 CS4025: Machine Translation l Background, how languages differ l MT Techniques l Controlled languages For more.
Natural Language Processing Expectation Maximization.
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2005 Lecture 1 21 July 2005.
Machine Translation Dr. Radhika Mamidi. What is Machine Translation? A sub-field of computational linguistics It investigates the use of computer software.
Lemmatization Tagging LELA /20 Lemmatization Basic form of annotation involving identification of underlying lemmas (lexemes) of the words in.
Part II. Statistical NLP Advanced Artificial Intelligence Applications of HMMs and PCFGs in NLP Wolfram Burgard, Luc De Raedt, Bernhard Nebel, Lars Schmidt-Thieme.
English-Persian SMT Reza Saeedi 1 WTLAB Wednesday, May 25, 2011.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
1 Natural Language Processing Gholamreza Ghassem-Sani Fall 1383.
ASPECTS OF LINGUISTIC COMPETENCE 4 SEPT 09, 2013 – DAY 6 Brain & Language LING NSCI Harry Howard Tulane University.
Leksička semantika i pragmatika 3. predavanje. Machine Translation The Story of the Stone –=The Dream of the Red Chamber (Cao Xueqin 1792) Issues: (“Language.
Machine Translation  Machine translation is of one of the earliest uses of AI  Two approaches:  Traditional approach using grammars, rewrite rules,
Approaches to Machine Translation CSC 5930 Machine Translation Fall 2012 Dr. Tom Way.
Ideas for 100K Word Data Set for Human and Machine Learning Lori Levin Alon Lavie Jaime Carbonell Language Technologies Institute Carnegie Mellon University.
For Wednesday Read chapter 23 Homework: –Chapter 22, exercises 1,4, 7, and 14.
What you have learned and how you can use it : Grammars and Lexicons Parts I-III.
Rules, Movement, Ambiguity
Artificial Intelligence: Natural Language
CSE573 Autumn /23/98 Natural Language Processing Administrative –PS3 due today –PS4 out Wednesday, due Friday 3/13 (last day of class) special.
Auckland 2012Kilgarriff: NLP and Corpus Processing1 The contribution of NLP: corpus processing.
Jan 2005CSA4050 Machine Translation II1 CSA4050: Advanced Techniques in NLP Machine Translation II Direct MT Transfer MT Interlingual MT.
WORDS The term word is much more difficult to define in a technical sense, and like many other linguistic terms, there are often arguments about what exactly.
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
Sight Words.
CPSC 422, Lecture 27Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27 Nov, 16, 2015.
1 Introduction to Computational Linguistics Eleni Miltsakaki AUTH Spring 2006-Lecture 2.
Machine Translation Divergences: A Formal Description and Proposed Solution Bonnie J. Dorr University of Maryland Presented by: Soobia Afroz.
Knowledge Structure Vijay Meena ( ) Gaurav Meena ( )
Error Analysis of Two Types of Grammar for the purpose of Automatic Rule Refinement Ariadna Font Llitjós, Katharina Probst, Jaime Carbonell Language Technologies.
Word classes and part of speech tagging. Slide 1 Outline Why part of speech tagging? Word classes Tag sets and problem definition Automatic approaches.
October 10, 2003BLTS Kickoff Meeting1 Transfer with Strong Decoding Learning Module Transfer Rules {PP,4894} ;;Score: PP::PP [NP POSTP] -> [PREP.
Eliciting a corpus of word- aligned phrases for MT Lori Levin, Alon Lavie, Erik Peterson Language Technologies Institute Carnegie Mellon University.
Lecture 3: Rules of Evaluation CS150: Computer Science
A Simple English-to-Punjabi Translation System By : Shailendra Singh.
Jan 2012MT Architectures1 Human Language Technology Machine Translation Architectures Direct MT Transfer MT Interlingual MT.
Introduction to Machine Translation
Linguistics 1 Syntax Week 2 Lectures 3 & 4.
Changes in English 1 In this presentation we are going to look at the way other languages have influenced English and at the similarities and differences.
RECENT TRENDS IN SMT By M.Balamurugan, Phd Research Scholar,
Introduction to Machine Translation
CSCI 5832 Natural Language Processing
Lecture 9: Machine Translation (I) October 25, 2005 Dan Jurafsky
Machine Translation Nov 8, 2006
How to use a dictionary effectively
CSCI 5832 Natural Language Processing
Introduction to Machine Translation
Lecture 12: Machine Translation (II) November 4, 2004 Dan Jurafsky
Information Retrieval
CS249: Neural Language Model
Presentation transcript:

Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2007 Lecture 14b 24 August 2007

Lecture 1, 7/21/2005Natural Language Processing2 LING 180 SYMBSYS 138 Intro to Computer Speech and Language Processing Lecture 9: Machine Translation (I) November 7, 2006 Dan Jurafsky Thanks to Bonnie Dorr for some of these slides!!

Lecture 1, 7/21/2005Natural Language Processing3 Outline for MT Week  Intro and a little history  Language Similarities and Divergences  Three classic MT Approaches Transfer Interlingua Direct  Modern Statistical MT  Evaluation

Lecture 1, 7/21/2005Natural Language Processing4 What is MT?  Translating a text from one language to another automatically.

Lecture 1, 7/21/2005Natural Language Processing5 Machine Translation  dai yu zi zai chuang shang gan nian bao chai you ting jian chuang wai zhu shao xiang ye zhe shang, yu sheng xi li, qing han tou mu, bu jue you di xia lei lai.  Dai-yu alone on bed top think-of-with-gratitude Bao-chai again listen to window outside bamboo tip plantain leaf of on-top rain sound sigh drop clear cold penetrate curtain not feeling again fall down tears come  As she lay there alone, Dai-yu’s thoughts turned to Bao-chai… Then she listened to the insistent rustle of the rain on the bamboos and plantains outside her window. The coldness penetrated the curtains of her bed. Almost without noticing it she had begun to cry.

Lecture 1, 7/21/2005Natural Language Processing6 Machine Translation

Lecture 1, 7/21/2005Natural Language Processing7 Machine Translation  The Story of the Stone  =The Dream of the Red Chamber (Cao Xueqin 1792)  Issues: Word segmentation Sentence segmentation: 4 English sentences to 1 Chinese Grammatical differences  Chinese rarely marks tense:  As, turned to, had begun,  tou -> penetrated  Zero anaphora  No articles Stylistic and cultural differences  Bamboo tip plaintain leaf -> bamboos and plantains  Ma ‘curtain’ -> curtains of her bed  Rain sound sigh drop -> insistent rustle of the rain

Lecture 1, 7/21/2005Natural Language Processing8 Not just literature  Hansards: Canadian parliamentary proceeedings

Lecture 1, 7/21/2005Natural Language Processing9 What is MT not good for?  Really hard stuff Literature Natural spoken speech (meetings, court reporting)  Really important stuff Medical translation in hospitals, 911

Lecture 1, 7/21/2005Natural Language Processing10 What is MT good for?  Tasks for which a rough translation is fine Web pages,  Tasks for which MT can be post-edited MT as first pass “Computer-aided human translation  Tasks in sublanguage domains where high-quality MT is possible FAHQT

Lecture 1, 7/21/2005Natural Language Processing11 Sublanguage domain  Weather forecasting “Cloudy with a chance of showers today and Thursday” “Low tonight 4”  Can be modeling completely enough to use raw MT output  Word classes and semantic features like MONTH, PLACE, DIRECTION, TIME POINT

Lecture 1, 7/21/2005Natural Language Processing12 MT History  1946 Booth and Weaver discuss MT at Rockefeller foundation in New York;  idea of dictionary-based direct translation  1949 Weaver memorandum popularized idea  1952 all 18 MT researchers in world meet at MIT  1954 IBM/Georgetown Demo Russian-English MT  lots of labs take up MT

Lecture 1, 7/21/2005Natural Language Processing13 History of MT: Pessimism  1959/1960: Bar-Hillel “Report on the state of MT in US and GB” Argued FAHQT too hard (semantic ambiguity, etc) Should work on semi-automatic instead of automatic His argument Little John was looking for his toy box. Finally, he found it. The box was in the pen. John was very happy. Only human knowledge let’s us know that ‘playpens’ are bigger than boxes, but ‘writing pens’ are smaller His claim: we would have to encode all of human knowledge

Lecture 1, 7/21/2005Natural Language Processing14 History of MT: Pessimism  The ALPAC report Headed by John R. Pierce of Bell Labs Conclusions:  Supply of human translators exceeds demand  All the Soviet literature is already being translated  MT has been a failure: all current MT work had to be post-edited  Sponsored evaluations which showed that intelligibility and informativeness was worse than human translations Results:  MT research suffered  Funding loss  Number of research labs declined  Association for Machine Translation and Computational Linguistics dropped MT from its name

Lecture 1, 7/21/2005Natural Language Processing15 History of MT  1976 Meteo, weather forecasts from English to French  Systran (Babelfish) been used for 40 years  1970’s: European focus in MT; mainly ignored in US  1980’s ideas of using AI techniques in MT (KBMT, CMU)  1990’s Commercial MT systems Statistical MT Speech-to-speech translation

Lecture 1, 7/21/2005Natural Language Processing16 Language Similarities and Divergences  Some aspects of human language are universal or near- universal, others diverge greatly.  Typology: the study of systematic cross-linguistic similarities and differences  What are the dimensions along with human languages vary?

Lecture 1, 7/21/2005Natural Language Processing17 Morphological Variation  Isolating languages Cantonese, Vietnamese: each word generally has one morpheme  Vs. Polysynthetic languages Siberian Yupik (`Eskimo’): single word may have very many morphemes  Agglutinative languages Turkish: morphemes have clean boundaries  Vs. Fusion languages Russian: single affix may have many morphemes

Lecture 1, 7/21/2005Natural Language Processing18 Syntactic Variation  SVO (Subject-Verb-Object) languages English, German, French, Mandarin  SOV Languages Japanese, Hindi  VSO languages Irish, Classical Arabic  SVO lgs generally prepositions: to Yuriko  VSO lgs generally postpositions: Yuriko ni

Lecture 1, 7/21/2005Natural Language Processing19 Segmentation Variation  Not every writing system has word boundaries marked Chinese, Japanese, Thai, Vietnamese  Some languages tend to have sentences that are quite long, closer to English paragraphs than sentences: Modern Standard Arabic, Chinese

Lecture 1, 7/21/2005Natural Language Processing20 Inferential Load: cold vs. hot lgs  Some ‘cold’ languages require the hearer to do more “figuring out” of who the various actors in the various events are: Japanese, Chinese,  Other ‘hot’ languages are pretty explicit about saying who did what to whom. English

Lecture 1, 7/21/2005Natural Language Processing21 Inferential Load (2) All noun phrases in blue do not appear in Chinese text … But they are needed for a good translation

Lecture 1, 7/21/2005Natural Language Processing22 Lexical Divergences  Word to phrases: English “computer science” = French “informatique”  POS divergences Eng. ‘she likes/VERB to sing’ Ger. Sie singt gerne/ADV Eng ‘I’m hungry/ADJ Sp. ‘tengo hambre/NOUN

Lecture 1, 7/21/2005Natural Language Processing23 Lexical Divergences: Specificity  Grammatical constraints English has gender on pronouns, Mandarin not.  So translating “3rd person” from Chinese to English, need to figure out gender of the person!  Similarly from English “they” to French “ils/elles”  Semantic constraints English `brother’ Mandarin ‘gege’ (older) versus ‘didi’ (younger) English ‘wall’ German ‘Wand’ (inside) ‘Mauer’ (outside) German ‘Berg’ English ‘hill’ or ‘mountain’

Lecture 1, 7/21/2005Natural Language Processing24 Lexical Divergence: many-to-many

Lecture 1, 7/21/2005Natural Language Processing25 Lexical Divergence: lexical gaps  Japanese: no word for privacy  English: no word for Cantonese ‘haauseun’ or Japanese ‘oyakoko’ (something like `filial piety’)  English ‘cow’ versus ‘beef’, Cantonese ‘ngau’

Lecture 1, 7/21/2005Natural Language Processing26 Event-to-argument divergences  English The bottle floated out.  Spanish La botella salió flotando. The bottle exited floating  Verb-framed lg: mark direction of motion on verb Spanish, French, Arabic, Hebrew, Japanese, Tamil, Polynesian, Mayan, Bantu familiies  Satellite-framed lg: mark direction of motion on satellite Crawl out, float off, jump down, walk over to, run after Rest of Indo-European, Hungarian, Finnish, Chinese

Lecture 1, 7/21/2005Natural Language Processing27 Structural divergences  G: Wir treffen uns am Mittwoch  E: We’ll meet on Wednesday

Lecture 1, 7/21/2005Natural Language Processing28 Head Swapping  E: X swim across Y  S: X crucar Y nadando  E: I like to eat  G: Ich esse gern  E: I’d prefer vanilla  G: Mir wäre Vanille lieber

Lecture 1, 7/21/2005Natural Language Processing29 Thematic divergence  Y me gusto  I like Y  G: Mir fällt der Termin ein  E: I forget the date

Lecture 1, 7/21/2005Natural Language Processing30 Divergence counts from Bonnie Dorr  32% of sentences in UN Spanish/English Corpus (5K) Categorial X tener hambre Y have hunger 98% Conflational X dar pu ñ aladas a Z X stab Z 83% Structural X entrar en Y X enter Y 35% Head Swapping X cruzar Y nadando X swim across Y 8% Thematic X gustar a Y Y likes X 6%

Lecture 1, 7/21/2005Natural Language Processing31 MT on the web  Babelfish:  Google: ri&rls=en&q="1+taza+de+jugo"+%28zumo%29+de+n aranja+5+cucharadas+de+azucar+morena&btnG=Se arch ri&rls=en&q="1+taza+de+jugo"+%28zumo%29+de+n aranja+5+cucharadas+de+azucar+morena&btnG=Se arch

Lecture 1, 7/21/2005Natural Language Processing32 3 methods for MT  Direct  Transfer  Interlingua

Lecture 1, 7/21/2005Natural Language Processing33 Three MT Approaches: Direct, Transfer, Interlingual

Lecture 1, 7/21/2005Natural Language Processing34 Direct Translation  Proceed word-by-word through text  Translating each word  No intermediate structures except morphology  Knowledge is in the form of Huge bilingual dictionary word-to-word translation information  After word translation, can do simple reordering Adjective ordering English -> French/Spanish

Lecture 1, 7/21/2005Natural Language Processing35 Direct MT Dictionary entry

Lecture 1, 7/21/2005Natural Language Processing36 Direct MT

Lecture 1, 7/21/2005Natural Language Processing37 Problems with direct MT  German  Chinese

Lecture 1, 7/21/2005Natural Language Processing38 The Transfer Model  Idea: apply contrastive knowledge, i.e., knowledge about the difference between two languages  Steps: Analysis: Syntactically parse Source language Transfer: Rules to turn this parse into parse for Target language Generation: Generate Target sentence from parse tree

Lecture 1, 7/21/2005Natural Language Processing39 English to French  Generally English: Adjective Noun French: Noun Adjective Note: not always true  Route mauvaise ‘bad road, badly-paved road’  Mauvaise route ‘wrong road’)  But is a reasonable first approximation Rule:

Lecture 1, 7/21/2005Natural Language Processing40 Transfer rules

Lecture 1, 7/21/2005Natural Language Processing41 Lexical transfer  Transfer-based systems also need lexical transfer rules  Bilingual dictionary (like for direct MT)  English home:  German nach Hause (going home) Heim (home game) Heimat (homeland, home country) zu Hause (at home)  Can list “at home zu Hause”  Or do Word Sense Disambiguation

Lecture 1, 7/21/2005Natural Language Processing42 Systran: combining direct and transfer  Analysis Morphological analysis, POS tagging Chunking of NPs, PPs, phrases Shallow dependency parsing  Transfer Translation of idioms Word sense disambiguation Assigning prepositions based on governing verbs  Synthesis Apply rich bilingual dictionary Deal with reordering Morphological generation

Lecture 1, 7/21/2005Natural Language Processing43 Transfer: some problems  N 2 sets of transfer rules!  Grammar and lexicon full of language-specific stuff  Hard to build, hard to maintain

Lecture 1, 7/21/2005Natural Language Processing44 Interlingua  Intuition: Instead of lg-lg knowledge rules, use the meaning of the sentence to help  Steps: 1) translate source sentence into meaning representation 2) generate target sentence from meaning.

Lecture 1, 7/21/2005Natural Language Processing45 Interlingua for Mary did not slap the green witch

Lecture 1, 7/21/2005Natural Language Processing46 Interlingua  Idea is that some of the MT work that we need to do is part of other NLP tasks  E.g., disambiguating E:book S:‘libro’ from E:book S:‘reservar’  So we could have concepts like BOOKVOLUME and RESERVE and solve this problem once for each language

Lecture 1, 7/21/2005Natural Language Processing47 Direct MT: pros and cons (Bonnie Dorr)  Pros Fast Simple Cheap No translation rules hidden in lexicon  Cons Unreliable Not powerful Rule proliferation Requires lots of context Major restructuring after lexical substitution

Lecture 1, 7/21/2005Natural Language Processing48 Interlingual MT: pros and cons (B. Dorr)  Pros Avoids the N 2 problem Easier to write rules  Cons: Semantics is HARD Useful information lost (paraphrase)

Lecture 1, 7/21/2005Natural Language Processing49 The impossibility of translation  Hebrew “adonoi roi” for a culture without sheep or shepherds Something fluent and understandable, but not faithful:  “The Lord will look after me” Something faithful, but not fluent and nautral  “The Lord is for me like somebody who looks after animals with cotton-like hair”

Lecture 1, 7/21/2005Natural Language Processing50 What makes a good translation  Translators often talk about two factors we want to maximize:  Faithfulness or fidelity How close is the meaning of the translation to the meaning of the original (Even better: does the translation cause the reader to draw the same inferences as the original would have)  Fluency or naturalness How natural the translation is, just considering its fluency in the target language

Lecture 1, 7/21/2005Natural Language Processing51 Statistical MT: Faithfulness and Fluency formalized!  Best-translation of a source sentence S:  Developed by researchers who were originally in speech recognition at IBM  Called the IBM model

Lecture 1, 7/21/2005Natural Language Processing52 The IBM model  Hmm, those two factors might look familiar…  Yup, it’s Bayes rule:

Lecture 1, 7/21/2005Natural Language Processing53 More formally  Assume we are translating from a foreign language sentence F to an English sentence E: F = f 1, f 2, f 3,…, f m  We want to find the best English sentence E-hat = e 1, e 2, e 3,…, e n E-hat = argmax E P(E|F) = argmax E P(F|E)P(E)/P(F) = argmax E P(F|E)P(E) Translation ModelLanguage Model

Lecture 1, 7/21/2005Natural Language Processing54 The noisy channel model for MT

Lecture 1, 7/21/2005Natural Language Processing55 Fluency: P(T)  How to measure that this sentence That car was almost crash onto me  is less fluent than this one: That car almost hit me.  Answer: language models (N-grams!) For example P(hit|almost) > P(was|almost)  But can use any other more sophisticated model of grammar  Advantage: this is monolingual knowledge!

Lecture 1, 7/21/2005Natural Language Processing56 Faithfulness: P(S|T)  French: ça me plait [that me pleases]  English: that pleases me - most fluent I like it I’ll take that one  How to quantify this?  Intuition: degree to which words in one sentence are plausible translations of words in other sentence Product of probabilities that each word in target sentence would generate each word in source sentence.

Lecture 1, 7/21/2005Natural Language Processing57 Faithfulness P(S|T)  Need to know, for every target language word, probability of it mapping to every source language word.  How do we learn these probabilities?  Parallel texts! Lots of times we have two texts that are translations of each other If we knew which word in Source Text mapped to each word in Target Text, we could just count!

Lecture 1, 7/21/2005Natural Language Processing58 Faithfulness P(S|T)  Sentence alignment: Figuring out which source language sentence maps to which target language sentence  Word alignment Figuring out which source language word maps to which target language word

Lecture 1, 7/21/2005Natural Language Processing59 Big Point about Faithfulness and Fluency  Job of the faithfulness model P(S|T) is just to model “bag of words”; which words come from say English to Spanish.  P(S|T) doesn’t have to worry about internal facts about Spanish word order: that’s the job of P(T)  P(T) can do Bag generation: put the following words in order (from Kevin Knight) have programming a seen never I language better -actual the hashing is since not collision-free usually the is less perfectly the of somewhat capacity table

Lecture 1, 7/21/2005Natural Language Processing60 P(T) and bag generation: the answer  “Usually the actual capacity of the table is somewhat less, since the hashing is not collision-free”  How about: loves Mary John

Lecture 1, 7/21/2005Natural Language Processing61 Summary  Intro and a little history  Language Similarities and Divergences  Three classic MT Approaches Transfer Interlingua Direct  Modern Statistical MT  Evaluation