Stephan Vogel - Machine Translation1 Machine Translation Distortion Model Stephan Vogel Spring Semester 2011.

Slides:



Advertisements
Similar presentations
Statistical Machine Translation
Advertisements

All actions in PADDS begin with setting up the patient.
Memory Protection: Kernel and User Address Spaces  Background  Address binding  How memory protection is achieved.
SURFACE AREA and VOLUME with BASE 10 BLOCKS
CSI 3120, Implementing subprograms, page 1 Implementing subprograms The environment in block-structured languages The structure of the activation stack.
Statistical Machine Translation Part II: Word Alignments and EM Alexander Fraser Institute for Natural Language Processing University of Stuttgart
Statistical Machine Translation Part II: Word Alignments and EM Alexander Fraser ICL, U. Heidelberg CIS, LMU München Statistical Machine Translation.
IT253: Computer Organization Lecture 6: Assembly Language and MIPS: Programming Tonga Institute of Higher Education.
Statistical Machine Translation Part II – Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
$100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300.
Prof. Necula CS 164 Lecture 141 Run-time Environments Lecture 8.
DP-based Search Algorithms for Statistical Machine Translation My name: Mauricio Zuluaga Based on “Christoph Tillmann Presentation” and “ Word Reordering.
’ strict ’ strict ’ strict ’ lenient ‘ lenient ‘ lenient
1 Handling nested procedures Method 1 : static (access) links –Reference to the frame of the lexically enclosing procedure –Static chains of such links.
1 A Tree Sequence Alignment- based Tree-to-Tree Translation Model Authors: Min Zhang, Hongfei Jiang, Aiti Aw, et al. Reporter: 江欣倩 Professor: 陳嘉平.
CS 536 Spring Run-time organization Lecture 19.
A Tree-to-Tree Alignment- based Model for Statistical Machine Translation Authors: Min ZHANG, Hongfei JIANG, Ai Ti AW, Jun SUN, Sheng LI, Chew Lim TAN.
Hidden Markov Models Usman Roshan BNFO 601. Hidden Markov Models Alphabet of symbols: Set of states that emit symbols from the alphabet: Set of probabilities.
Run time vs. Compile time
1 Improving a Statistical MT System with Automatically Learned Rewrite Patterns Fei Xia and Michael McCord (Coling 2004) UW Machine Translation Reading.
Symmetric Probabilistic Alignment Jae Dong Kim Committee: Jaime G. Carbonell Ralf D. Brown Peter J. Jansen.
Application of RNNs to Language Processing Andrey Malinin, Shixiang Gu CUED Division F Speech Group.
MACHINE TRANSLATION AND MT TOOLS: GIZA++ AND MOSES -Nirdesh Chauhan.
Stephan Vogel - Machine Translation1 Machine Translation Word Alignment Stephan Vogel Spring Semester 2011.
Natural Language Processing Expectation Maximization.
Translation Model Parameters (adapted from notes from Philipp Koehn & Mary Hearne) 24 th March 2011 Dr. Declan Groves, CNGL, DCU
Stephan Vogel - Machine Translation1 Statistical Machine Translation Word Alignment Stephan Vogel MT Class Spring Semester 2011.
Chapter 7: Runtime Environment –Run time memory organization. We need to use memory to store: –code –static data (global variables) –dynamic data objects.
English-Persian SMT Reza Saeedi 1 WTLAB Wednesday, May 25, 2011.
Statistical Machine Translation Part IV – Log-Linear Models Alex Fraser Institute for Natural Language Processing University of Stuttgart Seminar:
$100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300.
Machine Translation Course 8 Diana Trandab ă ț Academic year:
Creating New Webpages Pioneer Preparatory School.
2010 Failures in Czech-English Phrase-Based MT 2010 Failures in Czech-English Phrase-Based MT Full text, acknowledgement and the list of references in.
1 Machine Translation MIRA and MBR Stephan Vogel Spring Semester 2011.
Chapter 6 Programming Languages (2) Introduction to CS 1 st Semester, 2015 Sanghyun Park.
Review: –What is an activation record? –What are the typical fields in an activation record? –What are the storage allocation strategies? Which program.
Machine Translation  Machine translation is of one of the earliest uses of AI  Two approaches:  Traditional approach using grammars, rewrite rules,
Reordering Model Using Syntactic Information of a Source Tree for Statistical Machine Translation Kei Hashimoto, Hirohumi Yamamoto, Hideo Okuma, Eiichiro.
Advanced MT Seminar Spring 2008 Instructors: Alon Lavie and Stephan Vogel.
Statistical Phrase Alignment Model Using Dependency Relation Probability Toshiaki Nakazawa and Sadao Kurohashi Kyoto University.
Web Design I Spring 2009 Kevin Cole Gallaudet University
Run-Time Storage Organization Compiler Design Lecture (03/23/98) Computer Science Rensselaer Polytechnic.
Topic #1: Introduction EE 456 – Compiling Techniques Prof. Carl Sable Fall 2003.
Runtime Organization (Chapter 6) 1 Course Overview PART I: overview material 1Introduction 2Language processors (tombstone diagrams, bootstrapping) 3Architecture.
Improving Named Entity Translation Combining Phonetic and Semantic Similarities Fei Huang, Stephan Vogel, Alex Waibel Language Technologies Institute School.
The Percent Proportion. To solve a proportion, multiply the numbers that are diagonal, divide by the one that’s left.
(Statistical) Approaches to Word Alignment
Introduction to ASIC flow and Verilog HDL
Statistical Machine Translation Part II: Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
Discriminative Modeling extraction Sets for Machine Translation Author John DeNero and Dan KleinUC Berkeley Presenter Justin Chiu.
ICS312 Introduction to Compilers Set 23. What is a Compiler? A compiler is software (a program) that translates a high-level programming language to machine.
Spring 2010 Lecture 4 Kristina Toutanova MSR & UW With slides borrowed from Philipp Koehn and Hwee Tou Ng LING 575: Seminar on statistical machine translation.
♣Multiplying Decimals♣
Statistical Machine Translation Part II: Word Alignments and EM
Start on the corner of your picture *(Not the corner of the paper)
Gridding BIG Paper One square at a time….
Flow Path Model of Superscalars
. - t !!l t. - 1f1f J - /\/\ - ' I __.
7.1 Fractions, Decimals and Percents
Bitwise Operators CS163 Fall 2018.
.. '.. ' 'i.., \. J'.....,....., ,., ,,.. '"'". ' · · f.. -··-·· '.,.. \...,., '.··.. ! f.f.
Machine Translation and MT tools: Giza++ and Moses
Gaussian Elimination.
Abstraction in Code Procedural Data NOT student created
Machine Translation and MT tools: Giza++ and Moses
Vocabulary Strategy Step 1 fiduciary Step 2 Step 3 Step 4
This is an Arithmagon….
QUESTION NUMBER 1.
Statistical Machine Translation Part VI – Phrase-based Decoding
Presentation transcript:

Stephan Vogel - Machine Translation1 Machine Translation Distortion Model Stephan Vogel Spring Semester 2011

Stephan Vogel - Machine Translation2 Recap: DM in Word Alignment Models lHMM alignment: Jump model lCan be conditioned on word classes lBalance between data and parameters in model lLarger corpora -> richer models F E 3 0 2

Stephan Vogel - Machine Translation3 Distance Model lDecoder typically generates target sequence sequentially, while jumping forth and back on source sentence lSimplest reordering model lCost of a reordering depends only on the distance of the reordering lDistribution can be estimated from alignment lOr just a Gaussian with mean 1 lOr log p( a j | a j-1, I) = a j – a j-1 i.e. reordering cost proportional to distance

Stephan Vogel - Machine Translation4 Lexicalized Reordering Models lInstead of conditioning on classes, condition on actual words lDifferent possibilities: lCondition on source words vs target words lCondition on words at start of jump (out-bound) vs words at landing point (in- bound) EE FF

Stephan Vogel - Machine Translation5 Block Distortion Model lGiven current block, look at links at the corners lTop: how did I come from previous phrase? lBottom: how do I continue to next phrase? F E Current Block Left Top Left Bottom Right Bottom Right Top Previous Block Next Block Current Block

Stephan Vogel - Machine Translation6 Block Distortion Model lTop-Left: prev-to-current = monotone F E Current Block Left Top Previous Block

Stephan Vogel - Machine Translation7 Block Distortion Model lTop-Right: prev-to-current = swap F E Current Block Right Top Previous Block

Stephan Vogel - Machine Translation8 Block Distortion Model lNeither top-left nor top-right: prev-to-current = disjoint F E Current Block Previous Block

Stephan Vogel - Machine Translation9 Block Distortion Model lBottom-Right: current-to-next = monotone F E Current Block Next Block

Stephan Vogel - Machine Translation10 Block Distortion Model lBottom-Left: current-to-next = swap F E Current Block Next Block

Stephan Vogel - Machine Translation11 Block Distortion Model lNeither bottom-Left nor bottom-right: current-to-next = disjoint F E Current Block Next Block

Stephan Vogel - Machine Translation12 Moses Code // orientation to previous E bool connectedLeftTop = isAligned( sentence, startF-1, startE-1 ); bool connectedRightTop = isAligned( sentence, endF+1, startE-1 ); if ( connectedLeftTop && !connectedRightTop) extractFileOrientation << "mono"; else if (!connectedLeftTop && connectedRightTop) extractFileOrientation << "swap"; else extractFileOrientation << "other"; // orientation to following E bool connectedLeftBottom = isAligned( sentence, startF-1, endE+1 ); bool connectedRightBottom = isAligned( sentence, endF+1, endE+1 ); if ( connectedLeftBottom && !connectedRightBottom) extractFileOrientation << " swap"; else if (!connectedLeftBottom && connectedRightBottom) extractFileOrientation << " mono"; else extractFileOrientation << " other";

Stephan Vogel - Machine Translation13 Block Distortion Model lFor each phrase pair 6 counts: 2 groups of 3 lFrom previous: monotone swap other lTo next: monotone swap other lNormalize for each group lWe do not model p( orientation | phase_pair_1, phrase_pair_2 ) lMany overlapping and embedded blocks lWould be too sparse lWe model p( orientation | phrase_pair, entering ) and p( orientation | phrase_pair, leaving ) lI.e. not really looking at the previous block, but only at the alignment link lFor each entry in the phrase table we have an entry in the distortion model

Stephan Vogel - Machine Translation14 Distortion Model Table acuerdo con el lugar de ||| according to the place of ||| acuerdo con nuestra información ||| according to our information ||| acuerdo de pesca con Marruecos ||| fisheries agreement with Morocco ||| acuerdo entre Israel y ||| agreement ||| acuerdo no porque sea bueno, ||| agreement not because it is good, ||| acuerdo sobre este punto ||| agreed on ||| acuerdos a largo plazo se iniciaron en ||| long-term arrangements began in ||| acuerdos globales, especialmente ||| global agreements - primarily ||| lMany entries … lPhrase pair seen only once lSimple smoothing

Stephan Vogel - Machine Translation15 Distance-based ITG Reordering Model lSimple ITG model had very weak reordering model lCondition it on size of blocks (subtrees) lCondition on distance (e.g. taken from HMM alignment) F E

Stephan Vogel - Machine Translation16 Summary lDistortion models in word alignment models lDecoders work on phrases -> distortion models or phrases lIn Moses: Block reordering (also called lexicalized) lConditioned on phrase pair lMonotone, swap, disjoint lAlternatives lBased on words at the boundaries lInbound/Outbound lEasy to have lexicalized distortion model for ITG