Yu-Chieh Wu Yue-Shi Lee Jie-Chi Yang National Central University, Taiwan Ming Chuan University, Taiwan Date: 2006/6/8 Reporter: Yu-Chieh Wu The Exploration.

Slides:



Advertisements
Similar presentations
Special Topics in Computer Science Advanced Topics in Information Retrieval Lecture 10: Natural Language Processing and IR. Syntax and structural disambiguation.
Advertisements

CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 2 (06/01/06) Prof. Pushpak Bhattacharyya IIT Bombay Part of Speech (PoS)
Learning and Inference for Hierarchically Split PCFGs Slav Petrov and Dan Klein.
May 2006CLINT-LN Parsing1 Computational Linguistics Introduction Approaches to Parsing.
R Yun-Nung Chen 資工碩一 陳縕儂 1 /39.  Non-projective Dependency Parsing using Spanning Tree Algorithms (HLT/EMNLP 2005)  Ryan McDonald, Fernando.
计算机科学与技术学院 Chinese Semantic Role Labeling with Dependency-driven Constituent Parse Tree Structure Hongling Wang, Bukang Wang Guodong Zhou NLP Lab, School.
Dependency Parsing Some slides are based on:
Linear Model Incorporating Feature Ranking for Chinese Documents Readability Gang Sun, Zhiwei Jiang, Qing Gu and Daoxu Chen State Key Laboratory for Novel.
Unsupervised Structure Prediction with Non-Parallel Multilingual Guidance July 27 EMNLP 2011 Shay B. Cohen Dipanjan Das Noah A. Smith Carnegie Mellon University.
GRAMMAR & PARSING (Syntactic Analysis) NLP- WEEK 4.
 They speak German  8.47 million of people live there.
Annotating language data Tomaž Erjavec Institut für Informationsverarbeitung Geisteswissenschaftliche Fakultät Karl-Franzens-Universität Graz Tomaž Erjavec.
Tagging with Hidden Markov Models. Viterbi Algorithm. Forward-backward algorithm Reading: Chap 6, Jurafsky & Martin Instructor: Paul Tarau, based on Rada.
Clients for XProtect VMS What’s new presentation
Parsing with CFG Ling 571 Fei Xia Week 2: 10/4-10/6/05.
Experiments with a Multilanguage Non-Projective Dependency Parser Giuseppe Attardi Dipartimento di Informatica Università di Pisa.
1 Noun Homograph Disambiguation Using Local Context in Large Text Corpora Marti A. Hearst Presented by: Heng Ji Mar. 29, 2004.
Learning Accurate, Compact, and Interpretable Tree Annotation Slav Petrov, Leon Barrett, Romain Thibaux, Dan Klein.
Features and Unification
CS 310 – Fall 2006 Pacific University CS310 Parsing with Context Free Grammars Today’s reference: Compilers: Principles, Techniques, and Tools by: Aho,
Växjö University Joakim Nivre Växjö University. 2 Who? Växjö University (800) School of Mathematics and Systems Engineering (120) Computer Science division.
1 Complementarity of Lexical and Simple Syntactic Features: The SyntaLex Approach to S ENSEVAL -3 Saif Mohammad Ted Pedersen University of Toronto, Toronto.
April 26, 2007Workshop on Treebanking, NAACL-HTL 2007 Rochester1 Treebanks and Parsing Jan Hajič Institute of Formal and Applied Linguistics School of.
Fall 2004 Lecture Notes #5 EECS 595 / LING 541 / SI 661 Natural Language Processing.
Dependency Parsing with Reference to Slovene, Spanish and Swedish Simon Corston-Oliver Anthony Aue Microsoft Research.
Treebanks as Training Data for Parsers Joakim Nivre Växjö University and Uppsala University
An SVMs Based Multi-lingual Dependency Parsing Yuchang CHENG, Masayuki ASAHARA and Yuji MATSUMOTO Nara Institute of Science and Technology.
Computational Methods to Vocalize Arabic Texts H. Safadi*, O. Al Dakkak** & N. Ghneim**
1 The Hidden Vector State Language Model Vidura Senevitratne, Steve Young Cambridge University Engineering Department.
1 Data-Driven Dependency Parsing. 2 Background: Natural Language Parsing Syntactic analysis String to (tree) structure He likes fish S NP VP NP VNPrn.
Corpus-based computational linguistics or computational corpus linguistics? Joakim Nivre Uppsala University Department of Linguistics and Philology.
1 A study on automatically extracted keywords in text categorization Authors:Anette Hulth and Be´ata B. Megyesi From:ACL 2006 Reporter: 陳永祥 Date:2007/10/16.
Experiments on Building Language Resources for Multi-Modal Dialogue Systems Goals identification of a methodology for adapting linguistic resources for.
DataLyzer software Training. Introduction The purpose of this PPT is to give you quick information on the functionality of DataLyzer and to guide you.
Inductive Dependency Parsing Joakim Nivre
Training dependency parsers by jointly optimizing multiple objectives Keith HallRyan McDonaldJason Katz- BrownMichael Ringgaard.
A Language Independent Method for Question Classification COLING 2004.
May 2006CLINT-LN Parsing1 Computational Linguistics Introduction Parsing with Context Free Grammars.
The ICT Statistical Machine Translation Systems for IWSLT 2007 Zhongjun He, Haitao Mi, Yang Liu, Devi Xiong, Weihua Luo, Yun Huang, Zhixiang Ren, Yajuan.
Interpreter CS 124 Reference: Gamma et al (“Gang-of-4”), Design Patterns Some material taken from:
Semi-supervised Training of Statistical Parsers CMSC Natural Language Processing January 26, 2006.
A Cascaded Finite-State Parser for German Michael Schiehlen Institut für Maschinelle Sprachverarbeitung Universität Stuttgart
A Systematic Exploration of the Feature Space for Relation Extraction Jing Jiang & ChengXiang Zhai Department of Computer Science University of Illinois,
Coarse-to-Fine Efficient Viterbi Parsing Nathan Bodenstab OGI RPE Presentation May 8, 2006.
1 Intelligente Analyse- und Informationssysteme Frank Reichartz, Hannes Korte & Gerhard Paass Fraunhofer IAIS, Sankt Augustin, Germany Dependency Tree.
Dependency Parser for Swedish Project for EDA171 by Jonas Pålsson Marcus Stamborg.
Statistical Decision-Tree Models for Parsing NLP lab, POSTECH 김 지 협.
Iceland 5/30-6/1/07 1 Parsing with Morphological Information for Treebank Construction Seth Kulick University of Pennsylvania.
Page 1 NAACL-HLT 2010 Los Angeles, CA Training Paradigms for Correcting Errors in Grammar and Usage Alla Rozovskaya and Dan Roth University of Illinois.
Weakly Supervised Training For Parsing Mandarin Broadcast Transcripts Wen Wang ICASSP 2008 Min-Hsuan Lai Department of Computer Science & Information Engineering.
NLP. Introduction to NLP The probabilities don’t depend on the specific words –E.g., give someone something (2 arguments) vs. see something (1 argument)
Supertagging CMSC Natural Language Processing January 31, 2006.
Automatic Grammar Induction and Parsing Free Text - Eric Brill Thur. POSTECH Dept. of Computer Science 심 준 혁.
Exploiting Reducibility in Unsupervised Dependency Parsing David Mareček and Zdeněk Žabokrtský Institute of Formal and Applied Linguistics Charles University.
MACHINE TRANSLATION PAPER 1 Daniel Montalvo, Chrysanthia Cheung-Lau, Jonny Wang CS159 Spring 2011.
Virtual Examples for Text Classification with Support Vector Machines Manabu Sassano Proceedings of the 2003 Conference on Emprical Methods in Natural.
Earley’s Algorithm J. Earley, "An efficient context-free parsing algorithm", Communications of the Association for Computing Machinery, 13:2:94-102, 1970."An.
Chap. 7, Syntax-Directed Compilation J. H. Wang Nov. 24, 2015.
Department of Computer Science The University of Texas at Austin USA Joint Entity and Relation Extraction using Card-Pyramid Parsing Rohit J. Kate Raymond.
Dependency Parsing Niranjan Balasubramanian March 24 th 2016 Credits: Many slides from: Michael Collins, Mausam, Chris Manning, COLNG 2014 Dependency Parsing.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Chinese Named Entity Recognition using Lexicalized HMMs.
An Adaptive Learning with an Application to Chinese Homophone Disambiguation from Yue-shi Lee International Journal of Computer Processing of Oriental.
Learning to Generate Complex Morphology for Machine Translation Einat Minkov †, Kristina Toutanova* and Hisami Suzuki* *Microsoft Research † Carnegie Mellon.
The University of Illinois System in the CoNLL-2013 Shared Task Alla RozovskayaKai-Wei ChangMark SammonsDan Roth Cognitive Computation Group University.
Language Identification and Part-of-Speech Tagging
Raymond J. Mooney University of Texas at Austin
CSC 594 Topics in AI – Natural Language Processing
David Mareček and Zdeněk Žabokrtský
Authorship Attribution Using Probabilistic Context-Free Grammars
COUNTRIES NATIONALITIES LANGUAGES.
Presentation transcript:

Yu-Chieh Wu Yue-Shi Lee Jie-Chi Yang National Central University, Taiwan Ming Chuan University, Taiwan Date: 2006/6/8 Reporter: Yu-Chieh Wu The Exploration of Deterministic and Efficient Dependency Parsing

Context Nivre ’ s method is a LINEAR-TIME parsing algorithm But it presumed the projective grammar relation for text –One solution is to applying the psuedo projectivization (Nivre and Nilson, 2005) In addition, non-projective words or roots were still kept in stack –Un-parsed words In multilingual scenario, some languages annotated labels for roots

In this paper Extend the time efficiency of the Nivre ’ s method DO NOT scan the word sequence multiple times –Perform the Niver ’ s algorithm –Only focused on the “ UN-PARSED ” words Efficiently label the roots

System Overview Nivre’s Parser Learner 1 Root Parser Learner 2 Post- Processor Learner 3 Un-Parsed Text Un-Parsed Text Parsed Text Parsed Text Un- Parsed Words Un- Parsed Words

Our solution is… To reduce the un-parsed rate –We performed both Forward parsing Backward parsing directions (usually better) To classify the remaining words in stacks –A root parser to identify the word is… Root (including root label) or not root To re-connect the non-projective words –A post-processor is used to re-construct the arcs Exhaustive from the sentence start –Regardless its children

Statistics of un-parsed rate (percentage) Un-Parsed Rate ForwardBackward Arabic Chinese Czech Danish Dutch German ForwardBackward Japanese Portugese Slovene Spanish200.5 Swedish Turkish2.54 Bulgarian AVG

Root Parser For each un-parsed words Word i-2 Bigram i-2 BiPOS i-2 Word i-1 Bigram i-1 BiPOS i-1 Word i Bigram i BiPOS i Word i+1 Bigram i+1 BiPOS i+1 Word i+2 Bigram i+2 BiPOS i+2 Child R Bigram BiPOS Child 0 Bigram BiPOS

Experimental Results A (New result) B (Old result) C (Maltparser) Statistic test A vs BB vs CA vs C Arabic NoYes Chinese YesNoYes Czech YesNoYes Danish No Dutch NoYes German YesNoYes Japanese YesNoYes Portugese NoYes Slovene NoYes Spanish YesNoYes Swedish Yes Turkish NoYes Bulgarian No AVG

Parsing performance of different grained POS tags and forward/backward parsing directions Parsing direction LA-Score POS grained LA-Score JaForward91.43 Forward Fine91.43 Backward85.75Coarse91.25 ArForward60.62 Backward Fine63.55 Backward63.55Coarse63.63 TuForward55.47 Forward Fine55.47 Backward55.59Coarse55.59

Conclusion In this paper, we investigate the how effect does the “ fast parser ” achieve The employed features were quite simple –Only C/F-POS tag and word form We extend the Nivre ’ s method –Root parser –Exhaustive post-processing

Questions ?

System Spec Ⅰ.Ⅰ. Parsing Algorithm: 1. Nivre's Algorithm (Nivre, 2003) 2. Root Parser 3. Exhaustive-based Post-processing Ⅱ.Ⅱ. Parser Characteristics: 1. Top-down + Bottom-up 2. Deterministic + Exhaustive 3. Labeling integrated 4. Projective Ⅲ.Ⅲ. Learner:SVM Light (Joachims, 1998) (1)One-versus-One (2)Linear Kernel Ⅳ.Ⅳ. Feature Set: 1. Lexical (Unigram/Bigram) 2. Fine-grained POS and Coarse grained BiCPOS Ⅴ.Ⅴ. Post-Processing: Another learner is used to re-recognize heads in stacks Ⅵ.Ⅵ. Additional/Extern al Resources: Non-Used