Dependency Trees and Machine Translation Vamshi Ambati Spring 2008 Adv MT Seminar 02 April 2008.

Slides:



Advertisements
Similar presentations
Statistical Machine Translation
Advertisements

The Learning Non-Isomorphic Tree Mappings for Machine Translation Jason Eisner - Johns Hopkins Univ. a b A B events of misinform wrongly report to-John.
Compiler Construction
Computational language: week 10 Lexical Knowledge Representation concluded Syntax-based computational language Sentence structure: syntax Context free.
Statistical Machine Translation Part II: Word Alignments and EM Alexander Fraser ICL, U. Heidelberg CIS, LMU München Statistical Machine Translation.
Statistical Machine Translation Part II – Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
Proceedings of the Conference on Intelligent Text Processing and Computational Linguistics (CICLing-2007) Learning for Semantic Parsing Advisor: Hsin-His.
Grammars, constituency and order A grammar describes the legal strings of a language in terms of constituency and order. For example, a grammar for a fragment.
Dependency Parsing Joakim Nivre. Dependency Grammar Old tradition in descriptive grammar Modern theroretical developments: –Structural syntax (Tesnière)
Dependency Parsing Some slides are based on:
1 Statistical NLP: Lecture 12 Probabilistic Context Free Grammars.
1 A Tree Sequence Alignment- based Tree-to-Tree Translation Model Authors: Min Zhang, Hongfei Jiang, Aiti Aw, et al. Reporter: 江欣倩 Professor: 陳嘉平.
A Tree-to-Tree Alignment- based Model for Statistical Machine Translation Authors: Min ZHANG, Hongfei JIANG, Ai Ti AW, Jun SUN, Sheng LI, Chew Lim TAN.
Novel Reordering Approaches in Phrase-Based Statistical Machine Translation S. Kanthak, D. Vilar, E. Matusov, R. Zens & H. Ney ACL Workshop on Building.
Statistical Phrase-Based Translation Authors: Koehn, Och, Marcu Presented by Albert Bertram Titles, charts, graphs, figures and tables were extracted from.
“Applying Morphology Generation Models to Machine Translation” By Kristina Toutanova, Hisami Suzuki, Achim Ruopp (Microsoft Research). UW Machine Translation.
1 Improving a Statistical MT System with Automatically Learned Rewrite Patterns Fei Xia and Michael McCord (Coling 2004) UW Machine Translation Reading.
Course Summary LING 575 Fei Xia 03/06/07. Outline Introduction to MT: 1 Major approaches –SMT: 3 –Transfer-based MT: 2 –Hybrid systems: 2 Other topics.
C SC 620 Advanced Topics in Natural Language Processing Lecture 24 4/22.
Semi-Automatic Learning of Transfer Rules for Machine Translation of Low-Density Languages Katharina Probst April 5, 2002.
Parsing SLP Chapter 13. 7/2/2015 Speech and Language Processing - Jurafsky and Martin 2 Outline  Parsing with CFGs  Bottom-up, top-down  CKY parsing.
Application of RNNs to Language Processing Andrey Malinin, Shixiang Gu CUED Division F Speech Group.
Probabilistic Parsing Ling 571 Fei Xia Week 5: 10/25-10/27/05.
Daniel Gildea (2003): Loosely Tree-Based Alignment for Machine Translation Linguistics 580 (Machine Translation) Scott Drellishak, 2/21/2006.
1 Statistical NLP: Lecture 13 Statistical Alignment and Machine Translation.
Microsoft Research Faculty Summit Robert Moore Principal Researcher Microsoft Research.
SI485i : NLP Set 9 Advanced PCFGs Some slides from Chris Manning.
Machine translation Context-based approach Lucia Otoyo.
1 CS546: Machine Learning and Natural Language Preparation to the Term Project: - Dependency Parsing - Dependency Representation for Semantic Role Labeling.
Tree Kernels for Parsing: (Collins & Duffy, 2001) Advanced Statistical Methods in NLP Ling 572 February 28, 2012.
Tree-adjoining grammar (TAG) is a grammar formalism defined by Aravind Joshi and introduced in Tree-adjoining grammars are somewhat similar to context-free.
Advanced Signal Processing 05/06 Reinisch Bernhard Statistical Machine Translation Phrase Based Model.
THE BIG PICTURE Basic Assumptions Linguistics is the empirical science that studies language (or linguistic behavior) Linguistics proposes theories (models)
Statistical Machine Translation Part V – Better Word Alignment, Morphology and Syntax Alexander Fraser CIS, LMU München Seminar: Open Source.
Scalable Inference and Training of Context- Rich Syntactic Translation Models Michel Galley, Jonathan Graehl, Keven Knight, Daniel Marcu, Steve DeNeefe.
Dependency Tree-to-Dependency Tree Machine Translation November 4, 2011 Presented by: Jeffrey Flanigan (CMU) Lori Levin, Jaime Carbonell In collaboration.
May 2006CLINT-LN Parsing1 Computational Linguistics Introduction Parsing with Context Free Grammars.
NUDT Machine Translation System for IWSLT2007 Presenter: Boxing Chen Authors: Wen-Han Chao & Zhou-Jun Li National University of Defense Technology, China.
Reordering Model Using Syntactic Information of a Source Tree for Statistical Machine Translation Kei Hashimoto, Hirohumi Yamamoto, Hideo Okuma, Eiichiro.
11 Chapter 14 Part 1 Statistical Parsing Based on slides by Ray Mooney.
What’s in a translation rule? Paper by Galley, Hopkins, Knight & Marcu Presentation By: Behrang Mohit.
1 Modeling Long Distance Dependence in Language: Topic Mixtures Versus Dynamic Cache Models Rukmini.M Iyer, Mari Ostendorf.
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
A non-contiguous Tree Sequence Alignment-based Model for Statistical Machine Translation Jun Sun ┼, Min Zhang ╪, Chew Lim Tan ┼ ┼╪
Supertagging CMSC Natural Language Processing January 31, 2006.
2003 (c) University of Pennsylvania1 Better MT Using Parallel Dependency Trees Yuan Ding University of Pennsylvania.
Natural Language Processing Lecture 15—10/15/2015 Jim Martin.
CPSC 422, Lecture 27Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27 Nov, 16, 2015.
1 Minimum Error Rate Training in Statistical Machine Translation Franz Josef Och Information Sciences Institute University of Southern California ACL 2003.
Dependency Parsing Parsing Algorithms Peng.Huang
Wei Lu, Hwee Tou Ng, Wee Sun Lee National University of Singapore
Statistical Machine Translation Part II: Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
Discriminative Modeling extraction Sets for Machine Translation Author John DeNero and Dan KleinUC Berkeley Presenter Justin Chiu.
Towards Syntactically Constrained Statistical Word Alignment Greg Hanneman : Advanced Machine Translation Seminar April 30, 2008.
A Syntax-Driven Bracketing Model for Phrase-Based Translation Deyi Xiong, et al. ACL 2009.
A Simple English-to-Punjabi Translation System By : Shailendra Singh.
1 Local Search for Optimal Permutations Jason Eisner and Roy Tromble with Very Large-Scale Neighborhoods in Machine Translation.
LING 575 Lecture 5 Kristina Toutanova MSR & UW April 27, 2010 With materials borrowed from Philip Koehn, Chris Quirk, David Chiang, Dekai Wu, Aria Haghighi.
Spring 2010 Lecture 4 Kristina Toutanova MSR & UW With slides borrowed from Philipp Koehn and Hwee Tou Ng LING 575: Seminar on statistical machine translation.
Dependency Parsing Niranjan Balasubramanian March 24 th 2016 Credits: Many slides from: Michael Collins, Mausam, Chris Manning, COLNG 2014 Dependency Parsing.
Natural Language Processing : Probabilistic Context Free Grammars Updated 8/07.
CSC 594 Topics in AI – Natural Language Processing
Approaches to Machine Translation
Eiji Aramaki* Sadao Kurohashi* * University of Tokyo
Approaches to Machine Translation
Statistical Machine Translation Papers from COLING 2004
A Path-based Transfer Model for Machine Translation
Compilers Principles, Techniques, & Tools Taught by Jing Zhang
Dekai Wu Presented by David Goss-Grubbs
Johns Hopkins 2003 Summer Workshop on Syntax and Statistical Machine Translation Chapters 5-8 Ethan Phelps-Goodman.
Presentation transcript:

Dependency Trees and Machine Translation Vamshi Ambati Spring 2008 Adv MT Seminar 02 April 2008

Today Introduction –Dependency formalism –Syntax in Machine Translation Dependency Tree based Machine Translation –By projection –By synchronous modeling Conclusion and Future

Today Introduction –Dependency formalism –Syntax in Machine Translation Dependency Tree based Machine Translation –By projection –By synchronous modeling Conclusion and Future

Dependency Trees John gave Mary an apple Phrase Structure Trees

Dependency Trees John:N gave:V Mary:N an:DT apple:N S VP NP Phrase Structure Trees: Labels

Dependency Trees John gave Mary an apple gave apple Head Percolation: - Usually done deterministically - Assuming one head per phrase*

Dependency Trees John Mary an gave apple

Dependency Trees JohnMarygaveanapple

Dependency Trees: Basics Child Dependent Modifier John gave SUBJ Parent Governor Head Modified (optional) The direction of arrows can be head-child or child-head (has to be mentioned)

Dependency Trees: Basics Properties –Every word has a single head/parent Except for the root –Completely connected tree –Acyclic If w i →w j then never w j → * w i Variants –Projective: Non-crossing between dependencies If w i ->w j, then for all k between i and j, either w k ->w i or w k - >w j holds –Non-Projective: Allow crossings between depdenencies

Projective dependency tree ounces Projectiveness: all the words between here finally depend on either on “ was ” or “. ” Example credit: Yuji Matsumoto, NAIST, Japan

Non-projective dependency tree Example from: R. McDonald and F. Pereira EACL, Direction of edges: from a parent to the children Note: Phrases thus extracted which are united by dependencies could be discontinuous

Dependency Grammar (DG) in the Grammar Formalism Timeline Panini (2600 years ago, India) recognised, distinguished and classified semantic, syntactic and morphological dependencies (Bharati, Natural Language Processing) The Arabic grammarians (1200 years ago, Iraq) recognised government and syntactic dependency structure, (The Foundations of Grammar - Owens) The Latin grammarians (800 years ago) recognised 'determination' and dependency structures. - Percival, "Reflections on the History of Dependency Notions“ Lucien Tesnie`re (1930s, France) developed a relatively formal and sophisticated theory of DG grammar for use in schools PSG, CCG etc were around the same time in early 20 th century Source: ELLSSI 2000 Tutorial on Dependency Grammars

Dependency Trees: some phenomenon DG has been widely accepted as a variant of PSG, but it is not strongly equivalent –Constituents are implicit in a DepTree and can be derived –Relations are explicit and can be labelled although optional –No explicit non-terminal nodes, which means no unary productions too –Can handle discontinuous phrases too Known problems with Coordination and Gerunds

Phrase structure vs Dependency Phrase structure suitable to languages with –rather fixed word order patterns –clear constituency structures –English etc Dependency structure suitable to languages with –greater freedom of word order –order is controlled more by pragmatic than by syntactic factors –Slavonic (Czech, Polish) and some Romance (Italian, spanish etc)

Today Introduction –Dependency formalism –Syntax in Machine Translation Dependency Tree based Machine Translation –By projection –By synchronous modeling Conclusion and Future

Phrasal SMT discussion Advantages: –Do not have to compose translations unnecessarily –Local re-ordering captured in phrases –Already specific to the domain and capture context locally Disadvantages: –Specificity and no generalization –Discontiguous phrases not considered –Global reordering –Estimation problems (long vs short phrases) –Can not model phenomenon across phrases Limitations: –Phrase sizes (how much before I run into out of memory?) –Corpus Availability makes it feasible only to certain language pairs

Syntax in MT: Many Representations WordLevel MT : No syntax SMT: Phrases / contiguous sequences SMT Hierarchical : Pseudo Syntax Syntax based SMT : Constituent Syntax based SMT: CCG Syntax based SMT: LFG Syntax based SMT: Dependency

Syntax in MT: Many ways of incorporation Pre-procesing –Reordering input –Reordered training corpus Translation models –Syntactically informed alignment models –Better distortion models Language Models –Syntactic language models –Syntax motivated models Post-processing –Nbest list reranking with syntactic information –Translation correction: Case marker/TAM correction –True casing etc? –Multi combinations with Syntactic backbones?

Syntax based SMT discussion Inversion Transduction Grammar (Wu ‘96) –Very constrained form of syntax : One non-terminal –Some expressive limitations –Not linguistically motivated –Effectively learns preferences for flip/no-flip Generative Tree to String (Yamada & Knight 2001) –Expressiveness (last week presentation) –No discontiguous phrases Multitext grammars (Melamed 2003) – Formalized, but MT work yet to be realized Hierarchical MT (Chang 2005) –Linguistic generalizations –Handles discontiguous phrases recursively –Estimation problems and Phrase table are increased even more –Across phrase boundary modeling

Syntax in MT and Dependency Trees SourceTarget Syntax SourceTarget Syntax Tree and Tree Tree and String SeSe SeSe SfSf Source side tree is provided Target side is obtained by projection Problem of Isomorphism between trees head-switching empty-dep ; extra-dep Source side tree is provided Target side is provided Ideally non-isomorphic trees should be modeled too

Today Introduction –Dependency formalism –Syntax in Machine Translation Dependency Tree based Machine Translation –By projection –By synchronous modeling Conclusion and Future

Dependency Tree based Machine Translation By projection –Fox 2002 –Dekang Lin 2004 –Quirk et al 2004, Quirk et al 2006, Menezes et al 2007 By synchronous modeling –Alshawi et al 2001 –Jason Eisner 2003 –Fox 2005 –Yuang Lin and Daniel Marcu 2004

Phrasal Cohesion and Statistical Machine Translation Heidi Fox, EMNLP 2002 English-French Corpus was used –En-Fr are similar For phrase structure trees - –Head Crossings involve head constituent of the phrase with its modifier spans –Modifier Crossings involve only spans of modifier constituents For dependency trees –Head Crossings means crossings of spans of child with its parent –Modifier crossings same as above Dependency structures show cohesive nature across translation

A Path-based Transfer model Dekang Lin 2004 Input –Word-aligned –Source parsed Syntax translation model –Set of paths in source tree –Extract connected target path Generalization of paths to POS Modeling –Relative likelihood –Smoothing factor for noise

A Path-based Transfer model Dekang Lin 2004 Decoding –Parse input and extract all paths, extract target paths –Find a set of transfer rules Cover the entire source tree Can be consistently merged –Lexicalized rule preferred –Future work? Word ordering is addressed –Transfer rules from same sen: follow order in sentence –Only one example of path: follow order in rule –Many examples: pick relative distance from head Highest probability –Dynamic Programming Min-set cover problem applied to trees

A Path-based Transfer model Dekang Lin 2004 Evaluation –English-French: 1.2M –Source parsed by Minipar –1755 test set –5 to 15 words long sentences –Compared to Koehn’s results from 2003 paper No Language Model or extra generation module –Order defined by paths is linear –Some heuristics to maintain linearity Generalization of paths (transfer rules) quadratic vs. exponential Direct Correspondence Approach (DCA) is violated when translation divergences exist Very Naïve notion of reordering and merge conflict resolution SystemBLEU IBM PBSMT Current0.2612

Dependency Treelet Translation Quirk et al ACL 2004,05,06 Project dependencies from source to target via word alignment –One-one: project dependency to aligned words –Many-one: nothing to do, as the projected is the head –One-many : project to right most, and rest are attached to it Reattachment of modifiers to lowest possible node that preserves target word order Treelet extraction –All subtrees on source until a particular limit, and the corresponding target fragment which is connected –MLE for scoring

Dependency Treelet Translation Quirk et al ACL 2004,05,06 tired and dogs men fatigues et chiens hommes tired men and dogs hommes et chiens fatigues et chiens hommes fatigues Treelet with missing roots

Dependency Treelet Translation Quirk et al 2004,05,06 Translation Model –Trained from the aligned projected corpus –Log-linear with feature functions Channel Model –Treelet Prob –Lexical Prob Order Model –Head relative –Swap model Target Model –Target language model –Bigram Agreement model (opt)

Dependency Treelet Translation Quirk et al ACL 2004,05,06 Decoding (Step by step) –Input is a dependency analyzed source Challenge is that left-right may not work when starting with a Tree –Obtain best target tree combining the models –Exhaustive search using DP Translate bottom up, from a given subtree (ITG) For each head node extract all matching treelets: x_i –For each uncovered subtrees extract all matching treelets: y_i »Try all insertions of y_i into slots in x_i »Ordering model ranks all the re-ordering possibilities for the modifiers

Dependency Treelet Translation Quirk et al ACL 2004,05,06 Decoding Optimizations –Duplicate translations check&reuse –Nbest list (only maintain top best candidates) –Early pruning before reordering (channel model) –Greedy reordering (pick best one and move on) –Variable n-best size (dynamically reduce ‘n’ with increasing uncovered subtrees) –Determinstic pruning of treelets based on MLE (allowing decoder to try more reorderings) A* decoding –Estimate the cost of an uncovered node reordering instead of computing it exactly –Heuristics for optimistic estimates for each of the models

Dependency Treelet Translation Quirk et al ACL 2004,05,06 Evaluation –Eng-French –1.5M parallel Microsoft technical documentation –NLPWIN parsed on Eng side –GIZA++ trained –Target LM: French side of parallel data –Tuned on 250 sens for MaxBLEU –Tested on 10K unseen –1 Reference

Improvements to Treelet Translation Minimal Translation Units (HLT NAACL 2005) –Bilingual n-gram channel model (Banchs et.al 2005) M = m1 = –Instead of conditioning on the surface adjacent MTU, they condition on Headwordchain Dependency Order Templates (ACL 2007) Improve Generality in Translation Learn un-lexicalised order templates Only use at runtime for restricting search space in reordering

Dependency Tree based Machine Translation By projection –Fox 2002 –Dekang Lin 2004 –Quirk et al 2004, Quirk et al 2006, Menezes et al 2007 By synchronous modeling –Alshawi et al 2001 –Jason Eisner 2003 –Yuang Lin and Daniel Marcu 2004 –Fox 2005

Learning Dependency Translation Models as Collections of Finite-State Head Transducers Alshwai et al 2001 Head transducers variant –Middle-out string transduction vs. left-right –Can be used in a hierarchical fashion, if you consider input/output for non-head transitions as ‘strings’ rather than ‘words’ Dependency transduction model Empty in/out May not always be a dependency model in conventional sense

Learning Dependency Translation Models as Collections of Finite-State Head Transducers Alshwai et al 2001 Training: Given unaligned bitext –Compute coocurrence statistics at wordlevel –Find a hierarchical synchronous alignment driven by cost function –Construct a set of head transducers that explain the alignment –Calculate the transition weights by MLE Decoding –Similar to CKY or Chart Parsing, but ‘middle-out’ –Given input, find the best applications of transducers –A derivation spanning entire input means it probably has found best dependencies for source & target –Else string together most probable partial hypothesis to form a tree –Pick the target tree with lowest score and read off the string

Learning Dependency Translation Models as Collections of Finite-State Head Transducers Alshwai et al 2001 Evaluation –Eng – Spanish (ATIS data – 13,966 train, 1185 test) –Eng – Jap (Speech transcribed data – 12,226 train, 3253 test) Discussion –Language agnostic, direction agnostic –Induced dependency tree may not be syntactically motivated, but suited to translation –Application of transducers is done locally, and so less context information –A single transducer tries to do everything, training may have sparsity problems

Learning non-isomorphic tree mappings for MT Jason Eisner 2003 Non-Isomorphism not just due to language divergences but free translation A version of Tree Substitution Grammar –To learn from unaligned non-isomorphic trees –A statistical model based generalized instead of linguistic minimalism –Expressive with empty string insertions –Formulate for both PSG and DG Translation model –Joint model P (T s,T t,A) Alignment Decoding Training –Factorization helps: Reconstruct all derivations for a tree by efficient ‘tree parsing’ algorithm for TSG EM as an efficient inside-outside training on all derivations Decoding –Chart Parsing to create a forest of derivations for input tree –Maximize over probability of derivations –1-best derivation parse is syntactic-alignment 1. Kids kiss Sam quite often 2. Lots of kids give kisses to Sam

Machine Translation Using Probabilistic Synchronous Dependency Insertion Grammars Ding and Marcu 2005 SDIG –Like STAG, STIG for phrase structures –Basic units are elementary trees –Handles non-isomorphism at sub-tree level Cross-lingual inconsistencies are handled if they appear within basic units –Crossing-dependency –Broken-dependency

Machine Translation Using Probabilistic Synchronous Dependency Insertion Grammars Ding and Marcu 2005 Induction of SDIG for MT as Synchronous hierarchical tree partitioning –Train IBM Mode 1 scores for bitext –For each category of Node, starting with NP - –Perform synchronous tree partitioning operations Compute Prob of word pair (ei,fi) where operation can be performed Heuristic functions (Graphical model) guide the partitioning

Machine Translation Using Probabilistic Synchronous Dependency Insertion Grammars Ding and Marcu 2005 Translation Decoding for MT –Translation is obtained by maximizing over all possible derivations of the source tree translation of the ‘elementary trees’ –Analogous to HMM (Emission and Transition probs with elementary trees) –Decoding is similar to a Viterbi-style algorithm on the tree Hooks –Augmenting corpus by singleton ETs from Model1 –Smoothing probabilities

Machine Translation Using Probabilistic Synchronous Dependency Insertion Grammars Ding and Marcu 2005 Evaluation –Chinese-English system –Dan Bikels parses for both Cn,En trained from Parallel treebanks –Test with 4 refs Compared with –GIZA trained –ISI Rewrite Decoder NIST increased 97% BLUE increased 27% Reordering ignored for now

Dependency Based Statistical MT Fox 2005 Czech-English parallel corpus (Penn TB and Prague TB) –Morphological process and tecto-grammatical conversion for Czech trees –No processing for English trees Alignment of subtrees via IBM Model4 scores –followed by structural modification of trees to suit alignment (KEEP,SPLIT,BUD…) Translation Model :

Dependency Based Statistical MT Fox 2005 Decoding –Bestfirst decoder –Process given Czech input to dependency tree and translate each node independently –For each node Choose head position Generate english POS seq Generate the feature list Perform structural mutations Syntax Language Model –Takes as input a forest of phrase structures –Invert decoder forest output (dep tree nodes) into phrase structures –Reordering is entirely left to LM Evaluation –Work in progress –Proposed to use BLEU

Today Introduction –Dependency formalism –Syntax in Machine Translation Dependency Tree based Machine Translation –By projection –By synchronous modeling Conclusion and Future

Conclusion The good - –Easy to work with –Cohesive during projection –Builds well on top of existing PBSMT (Effective combination of lexicalization and syntax) –Supports modeling a target even with crossing phrase boundaries –Gracefully degrade over new domains The bad – –Reordering is not crucial, but expensive –Lots of hooks for decoding –Generalization explodes space The not so good – –Current approaches require a dependency tree on source side and a strong model for the target side

What Next… 1 year –Better scoring and estimation in syntactic translation models –Improvement in Dependency trees parse quality directly translates? (Chris Quirk et al 2006) ? What about MST Parser etc? –Better Word-Alignment and effect on model –Incorporating labeled dependencies. Will it help? –Factored Dependency Tree based models –Approximate subtree matching and Parsing algorithms 3-5 years –Decoding Algorithms and the Target-Ordering problem –Discriminative approaches to MT are catching up. How can syntax be incorporated into such a framework –Better syntactic language models based on the dependency formalisms –Semantics in Translation (Are DepTrees the first step?) –Fusion of Dependency and Constituent approaches (LFG style) –Joint Modeling approaches (Eisner 03, Smith 06 QS Grammar) –Taking MT to other applications like Cross-lingual Retrieval and QA which already use DepFormalisms

Thanks to Lori Levin: For discussion on Dependency tree formalism Amr Ahmed: For discussion and separation of work Respective authors of the papers for some of the graphic images I liberally used in the slides

Questions Thanks

DG Variants Case Grammar (Anderson) Daughter-Dependency Theory (Hudson) Dependency Unification Grammar (Hellwig) Functional-Generative Description (Sgall) Lexicase (Starosta) Meaning-Text Model (Mel'cuk) Metataxis (Schubert) Unification Dependency Grammar (Maxwell) Constraint Dependency Grammar (Maruyama)

Motivation Questions 1. How is dependency analysis used in Syntax MT? How do the algorithms vary if only the source side of analysis is present? 2. How do the decoding and transfer phases adapt when using dependency analysis? What algorithms exist and what is the complexity analysis? 3. How does dependency based syntax incorporation in MT, compare with other grammar formalisms like the phrase structure grammar? 4. Is there a class of languages which yield better to dependency analysis vs. other analysis? 5. Dependency analysis being close to semantics, does it help MT produce better results?

Other Papers QuasiSynchronous Grammars for Soft Syntactic Projection David Smith and Jason Eisner 2007 Automatic Learning of Parallel Dependency Treelet Pairs Yuan Ding and Martha Palmer 2004 Dependency vs. Constituents for Tree-Based Alignment Dan Gildea 2003 My Compilation: – ulehttp://kathmandu.lti.cs.cmu.edu:8080/wiki/index.php/AMT:Sched ule