Using Syntax to Disambiguate Explicit Discourse Connectives in Text Source: ACL-IJCNLP 2009 Author: Emily Pitler and Ani Nenkova Reporter: Yong-Xiang Chen.

Slides:



Advertisements
Similar presentations
Topics to be covered: Paragraph construction * general to specific/identify the problem-provide a solution Paragraph issues * using definitions * internal.
Advertisements

Feature Forest Models for Syntactic Parsing Yusuke Miyao University of Tokyo.
School of something FACULTY OF OTHER School of Computing FACULTY OF ENGINEERING Chunking: Shallow Parsing Eric Atwell, Language Research Group.
Emily Pitler, Annie Louis, Ani Nenkova University of Pennsylvania.
Exploring the Effectiveness of Lexical Ontologies for Modeling Temporal Relations with Markov Logic Eun Y. Ha, Alok Baikadi, Carlyle Licata, Bradford Mott,
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 2 (06/01/06) Prof. Pushpak Bhattacharyya IIT Bombay Part of Speech (PoS)
Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F Le Chesnay, France Jason Baldridge.
Progress update Lin Ziheng. System overview 2 Components – Connective classifier Features from Pitler and Nenkova (2009): – Connective: because – Self.
Automatically Evaluating Text Coherence Using Discourse Relations Ziheng Lin, Hwee Tou Ng and Min-Yen Kan Department of Computer Science National University.
Dr. Abdullah S. Al-Dobaian1 Ch. 2: Phrase Structure Syntactic Structure (basic concepts) Syntactic Structure (basic concepts)  A tree diagram marks constituents.
Using prosody to avoid ambiguity: Effects of speaker awareness and referential context Snedeker and Trueswell (2003) Psych 526 Eun-Kyung Lee.
Computational Models of Discourse Analysis Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Automatic Identification of Cognates, False Friends, and Partial Cognates University of Ottawa, Canada University of Ottawa, Canada.
1 Discourse, coherence and anaphora resolution Lecture 16.
Semantic Role Labeling Abdul-Lateef Yussiff
Made with OpenOffice.org 1 Sentiment Classification using Word Sub-Sequences and Dependency Sub-Trees Pacific-Asia Knowledge Discovery and Data Mining.
April 26th, 2007 Workshop on Treebanking, HLT/NAACL, Rochester 1 Layering of Annotations in the Penn Discourse TreeBank (PDTB) Rashmi Prasad Institute.
Recognizing Implicit Discourse Relations in the Penn Discourse Treebank Ziheng Lin, Min-Yen Kan, and Hwee Tou Ng Department of Computer Science National.
Predicting Text Quality for Scientific Articles AAAI/SIGART-11 Doctoral Consortium Annie Louis : Louis A. and Nenkova A Automatically.
Part of speech (POS) tagging
Parsing the NEGRA corpus Greg Donaker June 14, 2006.
A Framework for Named Entity Recognition in the Open Domain Richard Evans Research Group in Computational Linguistics University of Wolverhampton UK
Finding Advertising Keywords on Web Pages Scott Wen-tau YihJoshua Goodman Microsoft Research Vitor R. Carvalho Carnegie Mellon University.
SI485i : NLP Set 9 Advanced PCFGs Some slides from Chris Manning.
Context Free Grammars Reading: Chap 12-13, Jurafsky & Martin This slide set was adapted from J. Martin, U. Colorado Instructor: Paul Tarau, based on Rada.
PFA Node Alignment Algorithm Consider the parse trees of a Chinese-English parallel pair of sentences.
ELN – Natural Language Processing Giuseppe Attardi
Language and Culture Prof. R. Hickey SoSe 2006 How language works
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
1 Data-Driven Dependency Parsing. 2 Background: Natural Language Parsing Syntactic analysis String to (tree) structure He likes fish S NP VP NP VNPrn.
Distributional Part-of-Speech Tagging Hinrich Schütze CSLI, Ventura Hall Stanford, CA , USA NLP Applications.
A Comparison of Features for Automatic Readability Assessment Lijun Feng 1 Matt Huenerfauth 1 Martin Jansche 2 No´emie Elhadad 3 1 City University of New.
Discovery of Manner Relations and their Applicability to Question Answering Roxana Girju 1,2, Manju Putcha 1, and Dan Moldovan 1 University of Texas at.
1 Statistical NLP: Lecture 9 Word Sense Disambiguation.
Automatic classification for implicit discourse relations Lin Ziheng.
Page 1 Probabilistic Parsing and Treebanks L545 Spring 2000.
Ideas for 100K Word Data Set for Human and Machine Learning Lori Levin Alon Lavie Jaime Carbonell Language Technologies Institute Carnegie Mellon University.
A Systematic Exploration of the Feature Space for Relation Extraction Jing Jiang & ChengXiang Zhai Department of Computer Science University of Illinois,
Minimally Supervised Event Causality Identification Quang Do, Yee Seng, and Dan Roth University of Illinois at Urbana-Champaign 1 EMNLP-2011.
CSKGOI'08 Commonsense Knowledge and Goal Oriented Interfaces.
CSA2050 Introduction to Computational Linguistics Parsing I.
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
Supertagging CMSC Natural Language Processing January 31, 2006.
Automatic recognition of discourse relations Lecture 3.
UWMS Data Mining Workshop Content Analysis: Automated Summarizing Prof. Marti Hearst SIMS 202, Lecture 16.
Natural Language Processing Lecture 15—10/15/2015 Jim Martin.
CPSC 422, Lecture 27Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27 Nov, 16, 2015.
Shallow Parsing for South Asian Languages -Himanshu Agrawal.
Automatic sense prediction for implicit discourse relations in text Emily Pitler, Annie Louis, Ani Nenkova University of Pennsylvania ACL 2009.
◦ Process of describing the structure of phrases and sentences Chapter 8 - Phrases and sentences: grammar1.
Labeling protein-protein interactions Barbara Rosario Marti Hearst Project overview The problem Identifying the interactions between proteins. Labeling.
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
Learning Event Durations from Event Descriptions Feng Pan, Rutu Mulkar, Jerry R. Hobbs University of Southern California ACL ’ 06.
Word Sense and Subjectivity (Coling/ACL 2006) Janyce Wiebe Rada Mihalcea University of Pittsburgh University of North Texas Acknowledgements: This slide.
Computational Models of Discourse Analysis Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Statistical NLP Winter 2009
Statistical NLP: Lecture 3
Authorship Attribution Using Probabilistic Context-Free Grammars
Erasmus University Rotterdam
Improving a Pipeline Architecture for Shallow Discourse Parsing
Structural relations Carnie 2013, chapter 4 Kofi K. Saah.
Statistical NLP: Lecture 9
CSCI 5832 Natural Language Processing
Natural Language - General
Linguistic Essentials
Preposition error correction using Graph Convolutional Networks
Statistical NLP : Lecture 9 Word Sense Disambiguation
Presentation transcript:

Using Syntax to Disambiguate Explicit Discourse Connectives in Text Source: ACL-IJCNLP 2009 Author: Emily Pitler and Ani Nenkova Reporter: Yong-Xiang Chen

Discourse connectives Words or phrases that explicitly signal the presence of a discourse relation such as –once –since –on the contrary Implicit relations –a discourse connective is absent and inferred by the reader –hard to identify automatically Explicit relations are much easier to predict, but…

Two types of ambiguity 1.Discourse or non-discourse usage –For example, ” once” a temporal discourse connective a simply a word meaning “formerly” 2.Some connectives are ambiguous in terms of the relation they mark –For example, ” since” serve as temporal connective serve as causal connective

Goal Explore the predictive power of syntactic features for both disambiguation tasks

Corpus and features Corpus: Penn Discourse Treebank (PDTB) –Each discourse connective is assigned a sense from a three-level hierarchy of senses –Annotates 40,600 discourse relations (the largest public resource ) 18,459 Explicit Relations –of 100 explicit discourse connectives 16,053 Implicit Relations Other relations Annotators were allowed to provide two senses for a given connective

Relation categories of discourse connective in PDTB This work consider only the top level categories –general enough to be annotated with high inter- annotator agreement 1.Expansion 擴展 ( 遞進 / 解證 ) one clause is elaborating information in the other 2.Comparison 對比 ( 並列 ) information in the two clauses is compared or contrasted 3.Contingency 情況 ( 因果 / 條件 ) one clause expresses the cause of the other 4.Temporal 循序 ( 承接 ) information in two clauses are related because of their timing

Syntactic features Syntax has not been used for discourse vs. non- discourse disambiguation –Syntax extensively used for dividing sentences into elementary discourse units Idea: Discourse connectives appear in specific syntactic contexts Four feature categories: –Self Category –Parent Category –Left Sibling Category –Right Sibling Category Parent Left SiblingSelfRight Sibling

Self Category The highest node in the tree which dominates the words in the connective –For single word connectives this might correspond to the POS tag of the word –For multi-word connectives Example cue phrase “in addition” Parsed as (PP (IN In) (NP (NN addition) )) –Preposition + Noun –the Self Category of the phrase is prepositional phrase

Parent Category The category of the immediate parent of the Self Category –Example: My favorite colors are blue and green –when “and” doesn’t has a discourse function the parent of “and” would be an NP (“blue and green”)

Left Sibling Category The syntactic category of the sibling immediately to the left of the Self Category –If the left sibling does not exist, this features takes the value “NONE” Self Category has a discourse function –while in example above, the left sibling of “and” is “NP” so doesn’t has a discourse function

Right Sibling Category The syntactic category of the sibling immediately to the right of the Self Category English is a right-branching language –the right sibling is often the dependent of the potential discourse connective If the connective string has a discourse function –this dependent will often be a clause (SBAR) –Example: “After I went to the store, I went home” “After May, I will go on vacation”

More features about the right sibling Example: –NASA won’t attempt a rescue; instead, it will try to predict whether any of the rubble will smash to the ground and where. –Although the syntactic category of “where” is SBAR, “and” doesn’t has a discourse function So include two additional features about the contents of the right sibling –Right Sibling Contains a VP –Right Sibling Contains a Trace This example is a wh-trace

Discourse vs. non-discourse usage only 11 PDTB connectives appear as a discourse connective more than 90% of the time –although, in turn, afterward, consequently, additionally, alternatively, whereas, on the contrary, if and when, lest, and on the one hand...on the other hand –while “or” only serves a discourse function 2.8% of the times it appears

Training and testing Positive examples: –explicit discourse connectives annotated in the PDTB Negative examples: –same strings in the PDTB texts that were not annotated as explicit connectives report results using a maximum entropy classifier 2 sections (0 and 1) of the PDTB were used for development of the features 21 sections (2-22) used for ten-fold cross-validation Baseline: the string of the connective –f-score=75.33% Accuracy=85.86%

Combinations of features Different connectives have different syntactic contexts 1.pair-wise interaction features For example: connective=also-RightSibling=SBAR 2.Adding interaction terms between pairs of syntactic features

Sense classification a few connectives are quite ambiguous –since : indicates Temporal or Contingency –Contingency and Temporal are the senses most often annotated together. do classification between the four senses for each explicit relation using a Naive Bayes classifier The connectives most often doubly annotated are –when –and –as

Results The human inter-annotator agreement on the top level sense class was also 94% –suggesting further improvements may not be possible

Error Analysis Temporal relations are the least frequent of the four senses(19% of the explicit relations) But more than half of the errors involve the Temporal class –most commonly confused pairing was Contingency relations > Temporal relations –making up 29% of errors

Conclusion Using a few syntactic features leads to state-of-the-art accuracy for discourse vs. non-discourse usage classification Syntactic features also helps sense class identification –already attained results at the level of human annotator agreement