CMSC 723 / LING 645: Intro to Computational Linguistics September 8, 2004: Dorr MT (continued), MT Evaluation Prof. Bonnie J. Dorr Dr. Christof Monz TA:

Slides:



Advertisements
Similar presentations
Machine Translation: Interlingual Methods Thanks to Les Sikos Bonnie J. Dorr, Eduard H. Hovy, Lori S. Levin.
Advertisements

The Application of Machine Translation in CADAL Huang Chen, Chen Haiying Zhejiang University Libraries, Hangzhou, China
The Chinese Room: Understanding and Correcting Machine Translation This work has been supported by NSF Grants IIS Solution: The Chinese Room Conclusions.
Language Divergences and Solutions Advanced Machine Translation Seminar Alison Alvarez.
Statistical Machine Translation Part II – Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
Arthur Chan Prepared for Advanced MT Seminar
ADGEN USC/ISI ADGEN: Advanced Generation for Question Answering Kevin Knight and Daniel Marcu USC/Information Sciences Institute.
SEARCHING QUESTION AND ANSWER ARCHIVES Dr. Jiwoon Jeon Presented by CHARANYA VENKATESH KUMAR.
Machine Translation- 4 Autumn 2008 Lecture Sep 2008.
Baselines for Recognizing Textual Entailment Ling 541 Final Project Terrence Szymanski.
Re-evaluating Bleu Alison Alvarez Machine Translation Seminar February 16, 2006.
Essential ?s: How do I say ‘it’ and ‘them’ in Spanish, and how do I refer to people who get things?
1 I256: Applied Natural Language Processing Marti Hearst Sept 13, 2006.
BLEU, Its Variants & Its Critics Arthur Chan Prepared for Advanced MT Seminar.
Orange: a Method for Evaluating Automatic Evaluation Metrics for Machine Translation Chin-Yew Lin & Franz Josef Och (presented by Bilmes) or Orange: a.
CSCI 5582 Artificial Intelligence
Statistical Phrase-Based Translation Authors: Koehn, Och, Marcu Presented by Albert Bertram Titles, charts, graphs, figures and tables were extracted from.
Natural Language Query Interface Mostafa Karkache & Bryce Wenninger.
Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang National Central University
Translation Divergence LING 580MT Fei Xia 1/10/06.
Breaking the Resource Bottleneck for Multilingual Parsing Rebecca Hwa, Philip Resnik and Amy Weinberg University of Maryland.
1 The Web as a Parallel Corpus  Parallel corpora are useful  Training data for statistical MT  Lexical correspondences for cross-lingual IR  Early.
Essential ?s: How do I say ‘it’ and ‘them’ in Spanish, and how do I refer to people who get things?
Natural Language Processing Prof: Jason Eisner Webpage: syllabus, announcements, slides, homeworks.
A Pattern Matching Method for Finding Noun and Proper Noun Translations from Noisy Parallel Corpora Benjamin Arai Computer Science and Engineering Department.
CMSC 723 / LING 645: Intro to Computational Linguistics January 28, 2004 Lecture 1 (Dorr): Overview, History, Goals, Problems, Techniques; Intro to MT.
Machine Translation- 5 Autumn 2008 Lecture Sep 2008.
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
Machine Translation Dr. Radhika Mamidi. What is Machine Translation? A sub-field of computational linguistics It investigates the use of computer software.
1 Ling 569: Introduction to Computational Linguistics Jason Eisner Johns Hopkins University Tu/Th 1:30-3:20 (also this Fri 1-5)
English-Persian SMT Reza Saeedi 1 WTLAB Wednesday, May 25, 2011.
Introduction to Machine Translation Mitch Marcus CIS 530 Some slides adapted from slides by John Hutchins, Bonnie Dorr, Martha Palmer.
An Integrated Approach for Arabic-English Named Entity Translation Hany Hassan IBM Cairo Technology Development Center Jeffrey Sorensen IBM T.J. Watson.
Invitation to Computer Science, Java Version, Second Edition.
METEOR-Ranking & M-BLEU: Flexible Matching & Parameter Tuning for MT Evaluation Alon Lavie and Abhaya Agarwal Language Technologies Institute Carnegie.
Arthur Chan Prepared for Advanced MT Seminar
CMSC 723 / LING 723: Computational Linguistics I September 5, 2007: Dorr Part I: MT (cont), MT Evaluation (J & M, 24) Part II: Reg. Expressions, FSAutomata.
SMT – Final thoughts Philipp Koehn USC/Information Sciences Institute USC/Computer Science Department School of Informatics University of Edinburgh Some.
1 LIN 1310B Introduction to Linguistics Prof: Nikolay Slavkov TA: Qinghua Tang CLASS 22, March 27, 2007.
Natural Language Based Reformulation Resource and Web Exploitation for Question Answering Ulf Hermjakob, Abdessamad Echihabi, Daniel Marcu University of.
Machine Translation  Machine translation is of one of the earliest uses of AI  Two approaches:  Traditional approach using grammars, rewrite rules,
PARSING David Kauchak CS159 – Spring 2011 some slides adapted from Ray Mooney.
An Investigation of Statistical Machine Translation (Spanish to English) Raghav Bashyal.
CS460/IT632 Natural Language Processing/Language Technology for the Web Guest Lecture (31/03/06) Prof. Niladri Chatterjee IIT Delhi Guest Lecture on Machine.
MT with an Interlingua Lori Levin April 13, 2009.
What you have learned and how you can use it : Grammars and Lexicons Parts I-III.
Iterative Translation Disambiguation for Cross Language Information Retrieval Christof Monz and Bonnie J. Dorr Institute for Advanced Computer Studies.
Chinese Word Segmentation Adaptation for Statistical Machine Translation Hailong Cao, Masao Utiyama and Eiichiro Sumita Language Translation Group NICT&ATR.
1 Introduction to Computational Linguistics Eleni Miltsakaki AUTH Spring 2006-Lecture 1.
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
1 Minimum Error Rate Training in Statistical Machine Translation Franz Josef Och Information Sciences Institute University of Southern California ACL 2003.
Machine Translation Divergences: A Formal Description and Proposed Solution Bonnie J. Dorr University of Maryland Presented by: Soobia Afroz.
Statistical Machine Translation Part II: Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
Pastra and Saggion, EACL 2003 Colouring Summaries BLEU Katerina Pastra and Horacio Saggion Department of Computer Science, Natural Language Processing.
Instance Discovery and Schema Matching With Applications to Biological Deep Web Data Integration Tantan Liu, Fan Wang, Gagan Agrawal {liut, wangfa,
Machine Translation Course 10 Diana Trandab ă ț
A Simple English-to-Punjabi Translation System By : Shailendra Singh.
Review: Review: Translating without in-domain corpus: Machine translation post-editing with online learning techniques Antonio L. Lagarda, Daniel Ortiz-Martínez,
Essential ?s: How do I refer to people who get things?
Knowledge and Information Retrieval Dr Nicholas Gibbins 32/4037.
Automatic methods of MT evaluation Lecture 18/03/2009 MODL5003 Principles and applications of machine translation Bogdan Babych.
NLP Midterm Solution #1 bilingual corpora –parallel corpus (document-aligned, sentence-aligned, word-aligned) (4) –comparable corpus (4) Source.
Ling 575: Machine Translation Yuval Marton Winter 2016 February 9: MT Evaluation Much of the materials was borrowed from course slides of Chris Callison-Burch.
Machine Translation Course 9
METEOR: Metric for Evaluation of Translation with Explicit Ordering An Improved Automatic Metric for MT Evaluation Alon Lavie Joint work with: Satanjeev.
Machine Translation Nov 8, 2006
CS246: Information Retrieval
Lecture 12: Machine Translation (II) November 4, 2004 Dan Jurafsky
A Path-based Transfer Model for Machine Translation
Presented By: Sparsh Gupta Anmol Popli Hammad Abdullah Ayyubi
Presentation transcript:

CMSC 723 / LING 645: Intro to Computational Linguistics September 8, 2004: Dorr MT (continued), MT Evaluation Prof. Bonnie J. Dorr Dr. Christof Monz TA: Adam Lee

MT Challenges: Ambiguity  Syntactic Ambiguity I saw the man on the hill with the telescope  Lexical Ambiguity E: book S: libro, reservar  Semantic Ambiguity –Homography: ball(E) = pelota, baile(S) –Polysemy: kill(E), matar, acabar (S) –Semantic granularity esperar(S) = wait, expect, hope (E) be(E) = ser, estar(S) fish(E) = pez, pescado(S)

MT Challenges: Divergences  Meaning of two translationally equivalent phrases is distributed differently in the two languages  Example: –English: [RUN INTO ROOM] –Spanish: [ENTER IN ROOM RUNNING]

Divergence Frequency  32% of sentences in UN Spanish/English Corpus (5K)  35% of sentences in TREC El Norte Corpus (19K)  Divergence Types –Categorial (X tener hambre  X have hunger) [98%] –Conflational (X dar puñaladas a Z  X stab Z) [83%] –Structural (X entrar en Y  X enter Y)[35%] –Head Swapping (X cruzar Y nadando  X swim across Y)[8%] –Thematic (X gustar a Y  Y like X)[6%]

Spanish/Arabic Divergences Divergence E/E’ (Spanish) E/E’ (Arabic) Categorial be jealous when he returns have jealousy [tener celos] upon his return [ﻋﻧﺩ ﺮﺠﻭﻋﻪ] Conflational float come again go floating [ir flotando] return [ﻋﺎﺪ] Structural enter the house seek enter in the house [entrar en la casa] search for [ﺒﺣﺙ ﻋﻦ] Head Swap run in do something quickly enter running [entrar corriendo] go-quickly in doing something [ﺍﺴﺭﻉ] Thematic I have a headache my-head hurts me [me duele la cabeza] — [Arg1 [V]]  [Arg1 [MotionV] Modifier(v)] “The boat floated’’  “The boat went floating’’

(using narrowly defined divergence detection rules) Language Detected Human Sample Corpus Confirmed Size Size Spanish – Total 11.1% 10.5% 19K 150K Arabic – Total % 1K 28K Automatic Divergence Detection

Application of Divergence Detection: Bilingual Alignment for MT  Word-level alignments of bilingual texts are an integral part of MT models  Divergences present a great challenge to the alignment task  Common divergence types can be found in multiple language pairs, systematically identified, and resolved

The Problem: Alignment & Projection I began to eat the fish Yo empecé a comer el pescado

Why is this a hard problem? I run into the room Yo entro en el cuarto corriendo

Divergences! English: [RUN INTO ROOM] Spanish: [ENTER IN ROOM RUNNING]

Our Goal: Improved Alignment & Projection  Induce higher interannotator agreement rate  Increase the number of aligned words  Decrease multiple alignments

DUSTer Approach: Divergence Unraveling I run into the roomE: I move-in running the room E:E: Yo entro en el cuarto corriendo S:

Word-Level Alignment (1): Test Setup run John into room John enter room running Ex: John ran into the room → John entered the room running  Divergence Detection: Categorize English sentences into one of 5 divergence types  Divergence Correction: Apply appropriate structural transformation [E → E]

Word-Level Alignment (2): Testing Impact of Divergence Correction  Human align English and foreign sentence  Compare inter-annotator agreement, unaligned units, multiple alignments

Word-Level Alignment Results  Inter-Annotator Agreement: – English-Spanish: agreement increased from 80.2% to 82.9% – English-Arabic: agreement increased from 69.7% to 75.1%  Number of aligned words: – English-Spanish: aligned words increased from 82.8% to 86% – English-Arabic: aligned words increased from 61.5% to 88.1%  Multiple Alignments: – English-Spanish: number of links went from 1.35 to 1.16 – English-Arabic: number of links increased from 1.48 to 1.72

Divergence Unraveling Conclusions  Divergence handling shows promise for improvement of automatic alignment  Conservative lower bound on divergence frequency  Effective solution: syntactic transformation of English  Validity of solution shown through alignment experiments

How do we evaluate MT?  Human-based Metrics –Semantic Invariance –Pragmatic Invariance –Lexical Invariance –Structural Invariance –Spatial Invariance –Fluency –Accuracy –“Do you get it?”  Automatic Metrics: Bleu

BiLingual Evaluation Understudy (BLEU —Papineni, 2001)  Automatic Technique, but ….  Requires the pre-existence of Human (Reference) Translations  Approach: –Produce corpus of high-quality human translations –Judge “closeness” numerically (word-error rate) –Compare n-gram matches between candidate translation and 1 or more reference translations

Bleu Comparison Chinese-English Translation Example: Candidate 1: It is a guide to action which ensures that the military always obeys the commands of the party. Candidate 2: It is to insure the troops forever hearing the activity guidebook that party direct. Reference 1: It is a guide to action that ensures that the military will forever heed Party commands. Reference 2: It is the guiding principle which guarantees the military forces always being under the command of the Party. Reference 3: It is the practical guide for the army always to heed the directions of the party.

How Do We Compute Bleu Scores?  Key Idea: A reference word should be considered exhausted after a matching candidate word is identified. For each word compute: (1) candidate word count (2) maximum ref count Add counts for each candidate word using the lower of the two numbers. Divide by number of candidate words..

Modified Unigram Precision: Candidate #1 Reference 1: It is a guide to action that ensures that the military will forever heed Party commands. Reference 2: It is the guiding principle which guarantees the military forces always being under the command of the Party. Reference 3: It is the practical guide for the army always to heed the directions of the party. It(1) is(1) a(1) guide(1) to(1) action(1) which(1) ensures(1) that(2) the(4) military(1) always(1) obeys(0) the commands(1) of(1) the party(1) What’s the answer?????? 17/1 8

Modified Unigram Precision: Candidate #2 It(1) is(1) to(1) insure(0) the(4) troops(0) forever(1) hearing(0) the activity(0) guidebook(0) that(2) party(1) direct(0) What’s the answer?????? 8/1 4 Reference 1: It is a guide to action that ensures that the military will forever heed Party commands. Reference 2: It is the guiding principle which guarantees the military forces always being under the command of the Party. Reference 3: It is the practical guide for the army always to heed the directions of the party.

Modified Bigram Precision: Candidate #1 It is(1) is a(1) a guide(1) guide to(1) to action(1) action which(0) which ensures(0) ensures that(1) that the(1) the military(1) military always(0) always obeys(0) obeys the(0) the commands(0) commands of(0) of the(1) the party(1) What’s the answer?????? 10/1 7 Reference 1: It is a guide to action that ensures that the military will forever heed Party commands. Reference 2: It is the guiding principle which guarantees the military forces always being under the command of the Party. Reference 3: It is the practical guide for the army always to heed the directions of the party.

Modified Bigram Precision: Candidate #2 Reference 1: It is a guide to action that ensures that the military will forever heed Party commands. Reference 2: It is the guiding principle which guarantees the military forces always being under the command of the Party. Reference 3: It is the practical guide for the army always to heed the directions of the party. It is(1) is to(0) to insure(0) insure the(0) the troops(0) troops forever(0) forever hearing(0) hearing the(0) the activity(0) activity guidebook(0) guidebook that(0) that party(0) party direct(0) What’s the answer?????? 1/1 3

Catching Cheaters Reference 1: The cat is on the mat Reference 2: There is a cat on the mat the(2) the the the(0) the(0) the(0) the(0) What’s the unigram answer? 2/7 What’s the bigram answer?0/7