MEMT: Multi-Engine Machine Translation Faculty: Alon Lavie, Jaime Carbonell Students and Staff: Gregory Hanneman, Justin Merrill (Shyamsundar Jayaraman,

Slides:



Advertisements
Similar presentations
The Application of Machine Translation in CADAL Huang Chen, Chen Haiying Zhejiang University Libraries, Hangzhou, China
Advertisements

The Chinese Room: Understanding and Correcting Machine Translation This work has been supported by NSF Grants IIS Solution: The Chinese Room Conclusions.
Statistical Machine Translation Part II: Word Alignments and EM Alexander Fraser ICL, U. Heidelberg CIS, LMU München Statistical Machine Translation.
Statistical Machine Translation Part II – Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
Proceedings of the Conference on Intelligent Text Processing and Computational Linguistics (CICLing-2007) Learning for Semantic Parsing Advisor: Hsin-His.
Confidence Estimation for Machine Translation J. Blatz et.al, Coling 04 SSLI MTRG 11/17/2004 Takahiro Shinozaki.
Novel Reordering Approaches in Phrase-Based Statistical Machine Translation S. Kanthak, D. Vilar, E. Matusov, R. Zens & H. Ney ACL Workshop on Building.
The current status of Chinese-English EBMT research -where are we now Joy, Ralf Brown, Robert Frederking, Erik Peterson Aug 2001.
“Applying Morphology Generation Models to Machine Translation” By Kristina Toutanova, Hisami Suzuki, Achim Ruopp (Microsoft Research). UW Machine Translation.
TIDES MT Workshop Review. Using Syntax?  ISI-small: –Cross-lingual parsing/decoding Input: Chinese sentence + English lattice built with all possible.
Symmetric Probabilistic Alignment Jae Dong Kim Committee: Jaime G. Carbonell Ralf D. Brown Peter J. Jansen.
Semi-Automatic Learning of Transfer Rules for Machine Translation of Low-Density Languages Katharina Probst April 5, 2002.
MT Summit VIII, Language Technologies Institute School of Computer Science Carnegie Mellon University Pre-processing of Bilingual Corpora for Mandarin-English.
Scalable Text Mining with Sparse Generative Models
1 Statistical NLP: Lecture 13 Statistical Alignment and Machine Translation.
Microsoft Research Faculty Summit Robert Moore Principal Researcher Microsoft Research.
An Automatic Segmentation Method Combined with Length Descending and String Frequency Statistics for Chinese Shaohua Jiang, Yanzhong Dang Institute of.
Multi-Style Language Model for Web Scale Information Retrieval Kuansan Wang, Xiaolong Li and Jianfeng Gao SIGIR 2010 Min-Hsuan Lai Department of Computer.
Natural Language Processing Lab Northeastern University, China Feiliang Ren EBMT Based on Finite Automata State Transfer Generation Feiliang Ren.
Machine translation Context-based approach Lucia Otoyo.
Probabilistic Model for Definitional Question Answering Kyoung-Soo Han, Young-In Song, and Hae-Chang Rim Korea University SIGIR 2006.
METEOR-Ranking & M-BLEU: Flexible Matching & Parameter Tuning for MT Evaluation Alon Lavie and Abhaya Agarwal Language Technologies Institute Carnegie.
METEOR: Metric for Evaluation of Translation with Explicit Ordering An Automatic Metric for MT Evaluation with Improved Correlations with Human Judgments.
July 24, 2007GALE Update: Alon Lavie1 Statistical Transfer and MEMT Activities Multi-Engine Machine Translation –MEMT service within the cross-GALE IOD.
Transfer-based MT with Strong Decoding for a Miserly Data Scenario Alon Lavie Language Technologies Institute Carnegie Mellon University Joint work with:
MEMT: Multi-Engine Machine Translation Machine Translation Alon Lavie February 19, 2007.
Coping with Surprise: Multiple CMU MT Approaches Alon Lavie Lori Levin, Jaime Carbonell, Alex Waibel, Stephan Vogel, Ralf Brown, Robert Frederking Language.
Training dependency parsers by jointly optimizing multiple objectives Keith HallRyan McDonaldJason Katz- BrownMichael Ringgaard.
Recent Major MT Developments at CMU Briefing for Joe Olive February 5, 2008 Alon Lavie and Stephan Vogel Language Technologies Institute Carnegie Mellon.
NUDT Machine Translation System for IWSLT2007 Presenter: Boxing Chen Authors: Wen-Han Chao & Zhou-Jun Li National University of Defense Technology, China.
Advanced MT Seminar Spring 2008 Instructors: Alon Lavie and Stephan Vogel.
Approaches to Machine Translation CSC 5930 Machine Translation Fall 2012 Dr. Tom Way.
Transfer-based MT with Strong Decoding for a Miserly Data Scenario Alon Lavie Language Technologies Institute Carnegie Mellon University Joint work with:
MEMT: Multi-Engine Machine Translation Faculty: Alon Lavie, Robert Frederking, Ralf Brown, Jaime Carbonell Students: Shyamsundar Jayaraman, Satanjeev Banerjee.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
Statistical Machine Translation Part III – Phrase-based SMT / Decoding Alexander Fraser Institute for Natural Language Processing Universität Stuttgart.
Cluster-specific Named Entity Transliteration Fei Huang HLT/EMNLP 2005.
NRC Report Conclusion Tu Zhaopeng NIST06  The Portage System  For Chinese large-track entry, used simple, but carefully- tuned, phrase-based.
Improving Named Entity Translation Combining Phonetic and Semantic Similarities Fei Huang, Stephan Vogel, Alex Waibel Language Technologies Institute School.
A Trainable Transfer-based MT Approach for Languages with Limited Resources Alon Lavie Language Technologies Institute Carnegie Mellon University Joint.
A DYNAMIC APPROACH TO THE SELECTION OF HIGH ORDER N-GRAMS IN PHONOTACTIC LANGUAGE RECOGNITION Mikel Penagarikano, Amparo Varona, Luis Javier Rodriguez-
1 Minimum Error Rate Training in Statistical Machine Translation Franz Josef Och Information Sciences Institute University of Southern California ACL 2003.
Coping with Surprise: Multiple CMU MT Approaches Alon Lavie Lori Levin, Jaime Carbonell, Alex Waibel, Stephan Vogel, Ralf Brown, Robert Frederking Language.
A New Approach to Utterance Verification Based on Neighborhood Information in Model Space Author :Hui Jiang, Chin-Hui Lee Reporter : 陳燦輝.
A Trainable Transfer-based MT Approach for Languages with Limited Resources Alon Lavie Language Technologies Institute Carnegie Mellon University Joint.
The CMU Mill-RADD Project: Recent Activities and Results Alon Lavie Language Technologies Institute Carnegie Mellon University.
MEMT: Multi-Engine Machine Translation Guided by Explicit Word Matching Alon Lavie Language Technologies Institute Carnegie Mellon University Joint work.
Statistical Machine Translation Part II: Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
Discriminative Modeling extraction Sets for Machine Translation Author John DeNero and Dan KleinUC Berkeley Presenter Justin Chiu.
Large Vocabulary Data Driven MT: New Developments in the CMU SMT System Stephan Vogel, Alex Waibel Work done in collaboration with: Ying Zhang, Alicia.
July 24, 2007GALE Update: Alon Lavie1 Statistical Transfer and MEMT Activities Chinese-to-English Statistical Transfer MT system (Stat-XFER) –Developed.
October 10, 2003BLTS Kickoff Meeting1 Transfer with Strong Decoding Learning Module Transfer Rules {PP,4894} ;;Score: PP::PP [NP POSTP] -> [PREP.
CMU Statistical-XFER System Hybrid “rule-based”/statistical system Scaled up version of our XFER approach developed for low-resource languages Large-coverage.
MEMT: Multi-Engine Machine Translation Guided by Explicit Word Matching Faculty: Alon Lavie, Jaime Carbonell Students and Staff: Gregory Hanneman, Justin.
1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis.
Seed Generation and Seeded Version Space Learning Version 0.02 Katharina Probst Feb 28,2002.
CMU MilliRADD Small-MT Report TIDES PI Meeting 2002 The CMU MilliRADD Team: Jaime Carbonell, Lori Levin, Ralf Brown, Stephan Vogel, Alon Lavie, Kathrin.
MEMT: Multi-Engine Machine Translation Guided by Explicit Word Matching Alon Lavie Language Technologies Institute Carnegie Mellon University Joint work.
MEMT: Multi-Engine Machine Translation Faculty: Alon Lavie, Robert Frederking, Ralf Brown, Jaime Carbonell Students: Shyamsundar Jayaraman, Satanjeev Banerjee.
Discriminative n-gram language modeling Brian Roark, Murat Saraclar, Michael Collins Presented by Patty Liu.
LingWear Language Technology for the Information Warrior Alex Waibel, Lori Levin Alon Lavie, Robert Frederking Carnegie Mellon University.
1 Minimum Bayes-risk Methods in Automatic Speech Recognition Vaibhava Geol And William Byrne IBM ; Johns Hopkins University 2003 by CRC Press LLC 2005/4/26.
Multi-Engine Machine Translation
METEOR: Metric for Evaluation of Translation with Explicit Ordering An Improved Automatic Metric for MT Evaluation Alon Lavie Joint work with: Satanjeev.
Monoligual Semantic Text Alignment and its Applications in Machine Translation Alon Lavie March 29, 2012.
Translation Error Rate Metric
Statistical Machine Translation Part III – Phrase-based SMT / Decoding
Statistical Machine Translation Papers from COLING 2004
AMTEXT: Extraction-based MT for Arabic
Presenter : Jen-Wei Kuo
Presentation transcript:

MEMT: Multi-Engine Machine Translation Faculty: Alon Lavie, Jaime Carbonell Students and Staff: Gregory Hanneman, Justin Merrill (Shyamsundar Jayaraman, Satanjeev Banerjee)

March 10, 2005MEMT2 Goals and Approach Combine the output of multiple MT engines into a synthetic output that outperforms the originals in translation quality Synthetic combination of the originals, NOT selecting the best system Two main approaches: Approach-1: Merging of Lattice outputs + joint decoding –Each MT system produces a lattice of translation fragments, indexed based on source word positions –Lattices are merged into a single common lattice –Statistical MT decoder selects a translation “path” through the lattice Approach-2: Align best output from engines + new decoder –Each MT system produces a sentence translation output –Establish an explicit word matching between all words of the various MT engine outputs –“Decoding”: create a collection of synthetic combinations of the original strings based on matched words, target LM, and constraints + re-combination and pruning –Score resulting hypotheses and select a final output

March 10, 2005MEMT3 Synthetic Translation MEMT Idea: –Start with output sentences of the various MT engines –Explicitly align the words that are common between any pair of systems, and apply transitivity –Use the alignments as reinforcement and as indicators of possible locations for the words –Each engine has a “weight” that is used for the words that it contributes –Decoder searches for an optimal synthetic combination of words and phrases that optimizes a scoring function that combines the alignment weights and a LM score

March 10, 2005MEMT4 The Word-level Matcher Developed by Satanjeev Banerjee as a component in our METEOR Automatic MT Evaluation metric Finds maximal alignment match with minimal “crossing branches” Implementation: Clever search algorithm for best match using pruning of sub-optimal sub-solutions

March 10, 2005MEMT5 Matcher Example IBM: the sri lankan prime minister criticizes head of the country's ISI: The President of the Sri Lankan Prime Minister Criticized the President of the Country CMU: Lankan Prime Minister criticizes her country

March 10, 2005MEMT6 The MEMT Algorithm Algorithm builds collections of partial hypotheses of increasing length Partial hypotheses are extended by selecting the “next available” word from one of the original systems Sentences are assumed synchronous: –Each word is either aligned with another word or is an alternative of another word Extending a partial hypothesis with a word “pulls” and “uses” its aligned words with it, and marks its alternatives as “used” – “vectors” keep track of this Partial hypotheses are scored and ranked Pruning and re-combination Hypothesis can end if any original system proposes an end of sentence as next word

March 10, 2005MEMT7 The MEMT Algorithm Scoring: –Alignment score based on reinforcement from alignments of the words –LM score based on trigram LM –Sum logs of alignment score and LM score (equivalent to product of probabilities) –Select best scoring hypothesis based on: Total score (bias towards shorter hypotheses) Average score per word

March 10, 2005MEMT8 The MEMT Algorithm Parameters: –“lingering word” horizon: how long is a word allowed to linger when words following it have already been used? –“lookahead” horizon: how far ahead can we look for an alternative for a word that is not aligned? –“POS matching”: limit search for an alternative to only words of the same POS

March 10, 2005MEMT9 Example IBM: korea stands ready to allow visits to verify that it does not manufacture nuclear weapons ISI: North Korea Is Prepared to Allow Washington to Verify that It Does Not Make Nuclear Weapons CMU: North Korea prepared to allow Washington to the verification of that is to manufacture nuclear weapons Selected MEMT Sentence : north korea is prepared to allow washington to verify that it does not manufacture nuclear weapons ( )

March 10, 2005MEMT10 Example IBM: victims russians are one man and his wife and abusing their eight year old daughter plus a ( 11 and 7 years ) man and his wife and driver, egyptian nationality. : ISI: The victims were Russian man and his wife, daughter of the most from the age of eight years in addition to the young girls ) 11 7 years ( and a man and his wife and the bus driver Egyptian nationality. : CMU: the victims Cruz man who wife and daughter both critical of the eight years old addition to two Orient ( 11 ) 7 years ) woman, wife of bus drivers Egyptian nationality. : MEMT Sentence : Selected : the victims were russian man and his wife and daughter of the eight years from the age of a 11 and 7 years in addition to man and his wife and bus drivers egyptian nationality Oracle : the victims were russian man and wife and his daughter of the eight years old from the age of a 11 and 7 years in addition to the man and his wife and bus drivers egyptian nationality young girls

March 10, 2005MEMT11 Example IBM: the sri lankan prime minister criticizes head of the country's : ISI: The President of the Sri Lankan Prime Minister Criticized the President of the Country : CMU: Lankan Prime Minister criticizes her country: MEMT Sentence : Selected: the sri lankan prime minister criticizes president of the country Oracle: the sri lankan prime minister criticizes president of the country's

March 10, 2005MEMT12 Current System Initial development tests performed on TIDES 2003 Arabic-to-English MT data, using IBM, ISI and CMU SMT system output Further development tests performed on Arabic-to-English EBMT Apptek and SYSTRAN system output and on three Chinese-to-English COTS systems

March 10, 2005MEMT13 Experimental Results: Chinese-to-English SystemMETEOR Score Online Translator A.4917 Online Translator B.4859 Online Translator C.4910 Choosing best online translation.5381 MEMT.5301 Best hypothesis generated by MEMT.5840

March 10, 2005MEMT14 Experimental Results: Arabic-to-English SystemMETEOR Score Apptek.4241 EBMT.4231 Systran.4405 Choosing best online translation.4432 MEMT.5185 Best hypothesis generated by MEMT.5883

March 10, 2005MEMT15 Other Examples

March 10, 2005MEMT16 Architecture and Engineering Challenge: How do we construct an effective architecture for running MEMT within large- scale distributed projects? –Example: GALE Project –Multiple MT engines running at different locations –Input may be text or output of speech recognizers, Output may go downstream to other applications (IE, Summarization, TDT) Approach: Using IBM’s UIMA: Unstructured Information Management Architecture –Provides support for building robust processing “workflows” with heterogeneous components –Components act as “annotators” at the character level within documents

March 10, 2005MEMT17 UIMA-based MEMT MT engines and MEMT engine are set up as distributed servers: –Communication over socket connections –Sentence-by-sentence translation Java “wrappers” convert these into UIMA-style annotator components UIMA-based “workflows” implement a variety of a-synchronous tasks, with results stored in a common Annotations Database (ADB) –Translation workflows –MEMT workflow –Evaluation/scoring workflow

March 10, 2005MEMT18 UIMA-based MEMT: Examples Translation Workflow: –Retrieve document from ADB –“Annotate” document with translation annotator X –Write back new “annotation” into ADB MEMT Workflow: –Retrieve document translation annotations labeled by X, Y, Z from ADB –“Annotate” the document with a new MEMT annotation –Write back MEMT annotation into ADB

March 10, 2005MEMT19 Conclusions and Open Research Issues New sentence-level MEMT approach with promising performance Easy to run on both research and COTS systems UIMA-based architecture design for effective integration is large distributed systems/projects  GALE Main Open Research Issues: –Improvements to the underlying algorithm: better word alignments, “artificial” word alignments –Confidence scores at the sentence or word level –Decoding is still suboptimal Oracle scores show there is much room for improvement Need for additional discriminant features –Extend approach to Multi-Engine SR combination –Engineering issues: synchronization, human friendly workflows

March 10, 2005MEMT20

March 10, 2005MEMT21 Demo

March 10, 2005MEMT22 Approach-1: Lattice MEMT Approach: –Multiple MT systems produce a lattice of output segments –Create a “union” lattice of the various systems –Decode the joint lattice and select best synthetic output

March 10, 2005MEMT23 Approach-1: Lattice MEMT Lattice Decoder from CMU’s SMT: –Lattice arcs are scored uniformly using word-to-word translation probabilities, regardless of which engine produced the arc –Decoder searches for path that optimizes combination of Translation Model score and Language Model score –Decoder can also reorder words or phrases (up to 4 positions ahead)

March 10, 2005MEMT24 Initial Experiment: Hindi-to-English Systems Put together a scenario with “miserly” data resources: –Elicited Data corpus: phrases –Cleaned portion (top 12%) of LDC dictionary: ~2725 Hindi words (23612 translation pairs) –Manually acquired resources during the DARPA SLE: 500 manual bigram translations 72 manually written phrase transfer rules 105 manually written postposition rules 48 manually written time expression rules No additional parallel text!!

March 10, 2005MEMT25 Initial Experiment: Hindi-to-English Systems Tested on section of JHU provided data: 258 sentences with four reference translations –SMT system (stand-alone) –EBMT system (stand-alone) –XFER system (naïve decoding) –XFER system with “strong” decoder No grammar rules (baseline) Manually developed grammar rules Automatically learned grammar rules –XFER+SMT with strong decoder (MEMT)

March 10, 2005MEMT26 Results on JHU Test Set (very miserly training data) SystemBLEUM-BLEUNIST EBMT SMT XFER (naïve) man grammar XFER (strong) no grammar XFER (strong) learned grammar XFER (strong) man grammar XFER+SMT

March 10, 2005MEMT27 Effect of Reordering in the Decoder

March 10, 2005MEMT28 Further Experiments: Arabic-to-English Systems Combined: –CMU’s SMT system –CMU’s EBMT system –UMD rule-based system –(IBM didn’t work out) TM scores from CMU SMT system Built large new English LM Tested on TIDES 2003 Test set

March 10, 2005MEMT29 Arabic-to-English Systems Lattice MEMT Results: BLEUM-BLEUMETEOR UMD only.0335 [.0300,.0374].1099 [.1074,.1129].2356 [.2293,.3419] EBMT only.1090 [.1017,.1160].1861 [.1799,.1921].3666 [.3574,.3752] SMT only.2779 [.2782,.2886].3499 [.3412,.3582].5754 [.5649,.5855] EBMT+UMD.1206 [.1133,.1288].2069 [.2010,.2135].4061 [.3976,.4151] SMT+EBMT.2586 [.2477,.2702].3309 [.3222,.3403].5450 [.5360,.5545] SMT+UMD.2622 [.2519,.2724].3363 [.3281,.3446].5666 [.5575,.5764] SMT+UMD+ EBMT.2527 [.2426,.2640].3262 [.3181,.3349].5394 [.5290,.5504]

March 10, 2005MEMT30 Lattice MEMT Main Drawbacks: –Requires MT engines to provide lattice output  difficult to obtain! –Lattice output from all engines must be compatible: common indexing based on source word positions  difficult to standardize! –Common TM used for scoring edges may not work well for all engines –Decoding does not take into account any reinforcements from multiple engines proposing the same translation for any portion of the input

March 10, 2005MEMT31 Demonstration

March 10, 2005MEMT32 Experimental Results: Arabic-to-English SystemP/R/F1/Fmean Apptek.5137/.5336/.5235/.5316 EBMT.5710/.4781/.5204/.4860 Systran.4994/.5474/.5223/.5422 Choosing best online translation. MEMT.5383/.6212/.5768/.6118 Best hypothesis generated by MEMT.