Recognizing Implicit Discourse Relations in the Penn Discourse Treebank Ziheng Lin, Min-Yen Kan, and Hwee Tou Ng Department of Computer Science National.

Slides:



Advertisements
Similar presentations
Feature Forest Models for Syntactic Parsing Yusuke Miyao University of Tokyo.
Advertisements

Using Syntax to Disambiguate Explicit Discourse Connectives in Text Source: ACL-IJCNLP 2009 Author: Emily Pitler and Ani Nenkova Reporter: Yong-Xiang Chen.
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 2 (06/01/06) Prof. Pushpak Bhattacharyya IIT Bombay Part of Speech (PoS)
Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F Le Chesnay, France Jason Baldridge.
Progress update Lin Ziheng. System overview 2 Components – Connective classifier Features from Pitler and Nenkova (2009): – Connective: because – Self.
Polarity Analysis of Texts using Discourse Structure CIKM 2011 Bas Heerschop Erasmus University Rotterdam Frank Goossen Erasmus.
Automatically Evaluating Text Coherence Using Discourse Relations Ziheng Lin, Hwee Tou Ng and Min-Yen Kan Department of Computer Science National University.
LING 388: Language and Computers Sandiway Fong Lecture 2.
Exploiting Discourse Structure for Sentiment Analysis of Text OR 2013 Alexander Hogenboom In collaboration with Flavius Frasincar, Uzay Kaymak, and Franciska.
Playing the Telephone Game: Determining the Hierarchical Structure of Perspective and Speech Expressions Eric Breck and Claire Cardie Department of Computer.
April 26th, 2007 Workshop on Treebanking, HLT/NAACL, Rochester 1 Layering of Annotations in the Penn Discourse TreeBank (PDTB) Rashmi Prasad Institute.
Discourse Parsing in the Penn Discourse Treebank: Using Discourse Structures to Model Coherence and Improve User Tasks Ziheng Lin Ph.D. Thesis Proposal.
LEDIR : An Unsupervised Algorithm for Learning Directionality of Inference Rules Advisor: Hsin-His Chen Reporter: Chi-Hsin Yu Date: From EMNLP.
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
Predicting Text Quality for Scientific Articles AAAI/SIGART-11 Doctoral Consortium Annie Louis : Louis A. and Nenkova A Automatically.
1/13 Parsing III Probabilistic Parsing and Conclusions.
Automatic Classification of Semantic Relations between Facts and Opinions Koji Murakami, Eric Nichols, Junta Mizuno, Yotaro Watanabe, Hayato Goto, Megumi.
Introduction to CL Session 1: 7/08/2011. What is computational linguistics? Processing natural language text by computers  for practical applications.
Introduction to Syntax, with Part-of-Speech Tagging Owen Rambow September 17 & 19.
1 Natural Language Processing for the Web Prof. Kathleen McKeown 722 CEPSR, Office Hours: Wed, 1-2; Tues 4-5 TA: Yves Petinot 719 CEPSR,
Extracting Interest Tags from Twitter User Biographies Ying Ding, Jing Jiang School of Information Systems Singapore Management University AIRS 2014, Kuching,
Seven Lectures on Statistical Parsing Christopher Manning LSA Linguistic Institute 2007 LSA 354 Lecture 7.
Named Entity Recognition and the Stanford NER Software Jenny Rose Finkel Stanford University March 9, 2007.
Learning Table Extraction from Examples Ashwin Tengli, Yiming Yang and Nian Li Ma School of Computer Science Carnegie Mellon University Coling 04.
Summarization using Event Extraction Base System 01/12 KwangHee Park.
UAM CorpusTool: An Overview Debopam Das Discourse Research Group Department of Linguistics Simon Fraser University Feb 5, 2014.
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Tree Kernels for Parsing: (Collins & Duffy, 2001) Advanced Statistical Methods in NLP Ling 572 February 28, 2012.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
1 Data-Driven Dependency Parsing. 2 Background: Natural Language Parsing Syntactic analysis String to (tree) structure He likes fish S NP VP NP VNPrn.
A Survey of NLP Toolkits Jing Jiang Mar 8, /08/20072 Outline WordNet Statistics-based phrases POS taggers Parsers Chunkers (syntax-based phrases)
Distributional Part-of-Speech Tagging Hinrich Schütze CSLI, Ventura Hall Stanford, CA , USA NLP Applications.
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
Iterative Readability Computation for Domain-Specific Resources By Jin Zhao and Min-Yen Kan 11/06/2010.
10/12/2015CPSC503 Winter CPSC 503 Computational Linguistics Lecture 10 Giuseppe Carenini.
This work is supported by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior National Business Center contract number.
Describing Images using Inferred Visual Dependency Representations Authors : Desmond Elliot & Arjen P. de Vries Presentation of Paper by : Jantre Sanket.
Automatic classification for implicit discourse relations Lin Ziheng.
The ICT Statistical Machine Translation Systems for IWSLT 2007 Zhongjun He, Haitao Mi, Yang Liu, Devi Xiong, Weihua Luo, Yun Huang, Zhixiang Ren, Yajuan.
Recognizing Names in Biomedical Texts: a Machine Learning Approach GuoDong Zhou 1,*, Jie Zhang 1,2, Jian Su 1, Dan Shen 1,2 and ChewLim Tan 2 1 Institute.
A Bootstrapping Method for Building Subjectivity Lexicons for Languages with Scarce Resources Author: Carmen Banea, Rada Mihalcea, Janyce Wiebe Source:
A Cascaded Finite-State Parser for German Michael Schiehlen Institut für Maschinelle Sprachverarbeitung Universität Stuttgart
Opinion Holders in Opinion Text from Online Newspapers Youngho Kim, Yuchul Jung and Sung-Hyon Myaeng Reporter: Chia-Ying Lee Advisor: Prof. Hsin-Hsi Chen.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
A Systematic Exploration of the Feature Space for Relation Extraction Jing Jiang & ChengXiang Zhai Department of Computer Science University of Illinois,
Minimally Supervised Event Causality Identification Quang Do, Yee Seng, and Dan Roth University of Illinois at Urbana-Champaign 1 EMNLP-2011.
CSKGOI'08 Commonsense Knowledge and Goal Oriented Interfaces.
CSA2050 Introduction to Computational Linguistics Parsing I.
CPSC 503 Computational Linguistics
Supertagging CMSC Natural Language Processing January 31, 2006.
Automatic recognition of discourse relations Lecture 3.
CPSC 422, Lecture 27Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27 Nov, 16, 2015.
Support Vector Machines and Kernel Methods for Co-Reference Resolution 2007 Summer Workshop on Human Language Technology Center for Language and Speech.
Commonsense Reasoning in and over Natural Language Hugo Liu, Push Singh Media Laboratory of MIT The 8 th International Conference on Knowledge- Based Intelligent.
Automatic sense prediction for implicit discourse relations in text Emily Pitler, Annie Louis, Ani Nenkova University of Pennsylvania ACL 2009.
Using Wikipedia for Hierarchical Finer Categorization of Named Entities Aasish Pappu Language Technologies Institute Carnegie Mellon University PACLIC.
Exploiting Named Entity Taggers in a Second Language Thamar Solorio Computer Science Department National Institute of Astrophysics, Optics and Electronics.
Virtual Examples for Text Classification with Support Vector Machines Manabu Sassano Proceedings of the 2003 Conference on Emprical Methods in Natural.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Towards Semi-Automated Annotation for Prepositional Phrase Attachment Sara Rosenthal William J. Lipovsky Kathleen McKeown Kapil Thadani Jacob Andreas Columbia.
NLP. Parsing ( (S (NP-SBJ (NP (NNP Pierre) (NNP Vinken) ) (,,) (ADJP (NP (CD 61) (NNS years) ) (JJ old) ) (,,) ) (VP (MD will) (VP (VB join) (NP (DT.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
LING/C SC 581: Advanced Computational Linguistics Lecture Notes Feb 3 rd.
Parsing Natural Scenes and Natural Language with Recursive Neural Networks INTERNATIONAL CONFERENCE ON MACHINE LEARNING (ICML 2011) RICHARD SOCHER CLIFF.
The University of Illinois System in the CoNLL-2013 Shared Task Alla RozovskayaKai-Wei ChangMark SammonsDan Roth Cognitive Computation Group University.
PRESENTED BY: PEAR A BHUIYAN
Authorship Attribution Using Probabilistic Context-Free Grammars
Improving a Pipeline Architecture for Shallow Discourse Parsing
Artificial Intelligence 2004 Speech & Natural Language Processing
Presentation transcript:

Recognizing Implicit Discourse Relations in the Penn Discourse Treebank Ziheng Lin, Min-Yen Kan, and Hwee Tou Ng Department of Computer Science National University of Singapore EMNLP 2009

Outline Recognizing implicit discourse relations Related work Penn Discourse Treebank Model and features Experiments Discussion Conclusion

Implicit discourse relations When there is a discourse connective between two text spans, it is often easy to recognize the relation between the spans. – because, but, so.. It is difficult to recognize the discourse relations when there are no explicit textual cues. – I am in Singapore, but I live in the United States. – The main conference is over Wednesday. I am staying for EMNLP.

Background Recently, the Peen Discourse Treebank (PDTB) has been released, which features discourse level annotation on both explicit and implicit relations. With the recent release of the second version of this corpus, which provides a cleaner and more thorough implicit relation annotation. It is an opportunity to address this area of work.

Related work Generated implicit relation annotation – Marcu & Echihabi (2002) – Saito et al. (2006) Pitler et al. (2009) – Implicit relation classification on the second version of the PDTB. – Using several linguistically informed features, such as word polarity, verb classes, and word pairs. – Four-class classification (first level of relations)

Penn Discourse Treebank PDTB is a discourse level annotation over the one million word Wall Street Journal corpus. Annotations for Explicit and implicit discourse relations are provided. – In any case, the brokerage firms are clearly moving faster to create new ads than they did in the fall of But it remains to be seen whether their ads will be any more effective. – “A lot of investor confidence comes from the fact that they can speak to us,” he says. [so] “To maintain that dialogue is absolutely crucial.”

Methodology The implicit relation classifier is built using supervised learning on a maximum entropy (MaxEnt) classifier. Represent the annotated argument pairs into binary feature vectors suitable for training. Four classes of features as they represent a wide range of information. – Contextual – Constituent parse – Dependency parse – Lexical

Contextual features For each relation curr we want to classify, we use the previous relation prev and the next relation next as evidence to fire six binary features.

Constituent parse features Build the syntactic tree and collect all production rules from the parse tree, with function tags (e.g. SBJ) and POS tags. – S -> NP VP, NP -> PRP, PRP -> “We”

Dependency parse features Produce a dependency tree for each argument using Stanford dependency parser. The dependency trees encode additional information at the word level that is not explicit present in the constituent trees. – “had” <- nsubj dobj – “problems” <- det nn advmod – “at” <- dep

Lexical features Word pairs extracted from the respective text spans are a good signal of the discourse relation between arguments. Stemmed and collected all word pairs from Arg1 and Arg2, i.e., all (w i, w j ) where w i is a word from Arg1 and w j is a word from Arg2.

Feature selection Used a frequency cutoff of 5 to remove infrequent features. – 11,113 production rules. – 5,031 dependency rules. – 105,783 word pairs. Rank the features with mutual information (MI).

Experiment setup OpenNLP MaxEnt package PDTB Sections 2-21 as the training set and Section 23 as the test set. Baseline: the majority class Cause. – Accuracy of 26.1% on the test set.

Efficacy of the different feature classes

Performance for the 11 relations

Why are implicit discourse relations difficult to recognize? Ambiguity Inference Context – Possibly need to read the whole article and understand what was happening. World knowledge – Sometimes need world knowledge to understand the arguments and hence to interpret the relation. Implicit discourse relation classification needs deeper semantic representations, more robust system design, and access to more external knowledge.

Ambiguity Contrast and Conjunction in the PDTB annotation are very similar to each other. – In the third quarter, AMR said, net fell to $137 million, or $2.16 a share, from $150.3 million, or $2.50 a share. [while] Revenue rose 17% to $2.73 billion from $2.33 billion a year earlier. (Contrast) – Dow’s third quarter net fell to $589 million, or $3.29 a share, from $632 million, or $3.36 a share, a year ago. [while] Sales in the latest quarter rose 2% to $4.25 billion from $4.15 billion a year earlier. (Conjunction) – Sales surged 40% to billion yen from billion. [meanwhile] Net income rose 11% to billion yen from billion. (Conjunction; Contrast)

Inference Sometimes inference and a knowledge base are required to resolve the relation type. – “I had calls all night long from the States,” he said. [in fact] I was woken up every hour – 1:30, 2:30, 3:30, 4:30.” (Restatement) – I had calls all night long -> I was woken up every hour.

Conclusion Implement an implicit discourse relation classifier and showed initial results on the recently released Penn Discourse Treebank. The features include the modeling of the context of relations, features extracted from constituent parse trees and dependency parse trees, and word pair features. The classifier achieves an accuracy of 40.2%, a 14.1% improvement over the baseline. Conducted a data analysis and discussed four challenges that need to be addressed.