NYU Coreference CSCI-GA.2591 Ralph Grishman.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Multi-Document Person Name Resolution Michael Ben Fleischman (MIT), Eduard Hovy (USC) From Proceedings of ACL-42 Reference Resolution workshop 2004.
Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F Le Chesnay, France Jason Baldridge.
A Machine Learning Approach to Coreference Resolution of Noun Phrases By W.M.Soon, H.T.Ng, D.C.Y.Lim Presented by Iman Sen.
Overview of the TAC2013 Knowledge Base Population Evaluation: English Slot Filling Mihai Surdeanu with a lot help from: Hoa Dang, Joe Ellis, Heng Ji, and.
A Corpus for Cross- Document Co-Reference D. Day 1, J. Hitzeman 1, M. Wick 2, K. Crouch 1 and M. Poesio 3 1 The MITRE Corporation 2 University of Massachusetts,
The Impact of Task and Corpus on Event Extraction Systems Ralph Grishman New York University Malta, May 2010 NYU.
Processing of large document collections Part 6 (Text summarization: discourse- based approaches) Helena Ahonen-Myka Spring 2006.
Easy-First Coreference Resolution Veselin Stoyanov and Jason Eisner Johns Hopkins University.
Reference Resolution #1 CSCI-GA.2590 Ralph Grishman NYU.
Ang Sun Ralph Grishman Wei Xu Bonan Min November 15, 2011 TAC 2011 Workshop Gaithersburg, Maryland USA.
Clustering… in General In vector space, clusters are vectors found within  of a cluster vector, with different techniques for determining the cluster.
Event Extraction: Learning from Corpora Prepared by Ralph Grishman Based on research and slides by Roman Yangarber NYU.
CS 4705 Lecture 21 Algorithms for Reference Resolution.
Using Information Extraction for Question Answering Done by Rani Qumsiyeh.
Lance Ramshaw (with Ralph Weischedel) BBN. 2 Ontobank Coreference Part of the multi-site Ontobank effort –Intended to combine with word-sense and propositional.
Supervised models for coreference resolution Altaf Rahman and Vincent Ng Human Language Technology Research Institute University of Texas at Dallas 1.
Improving Machine Learning Approaches to Coreference Resolution Vincent Ng and Claire Cardie Cornell Univ. ACL 2002 slides prepared by Ralph Grishman.
Reference Resolution CSCI-GA.2590 Ralph Grishman NYU.
A Discriminative Latent Variable Model for Online Clustering Rajhans Samdani, Kai-Wei Chang, Dan Roth Department of Computer Science University of Illinois.
A Global Relaxation Labeling Approach to Coreference Resolution Coling 2010 Emili Sapena, Llu´ıs Padr´o and Jordi Turmo TALP Research Center Universitat.
A Light-weight Approach to Coreference Resolution for Named Entities in Text Marin Dimitrov Ontotext Lab, Sirma AI Kalina Bontcheva, Hamish Cunningham,
Andreea Bodnari, 1 Peter Szolovits, 1 Ozlem Uzuner 2 1 MIT, CSAIL, Cambridge, MA, USA 2 Department of Information Studies, University at Albany SUNY, Albany,
C OLLECTIVE ANNOTATION OF WIKIPEDIA ENTITIES IN WEB TEXT - Presented by Avinash S Bharadwaj ( )
Learning Structure in Bayes Nets (Typically also learn CPTs here) Given the set of random variables (features), the space of all possible networks.
Reyyan Yeniterzi Weakly-Supervised Discovery of Named Entities Using Web Search Queries Marius Pasca Google CIKM 2007.
Illinois-Coref: The UI System in the CoNLL-2012 Shared Task Kai-Wei Chang, Rajhans Samdani, Alla Rozovskaya, Mark Sammons, and Dan Roth Supported by ARL,
JHU/CLSP/WS07/ELERFED Scoring Metrics for IDC, CDC and EDC David Day ELERFED JHU Workshop July 18, 2007.
Efficiently Computed Lexical Chains As an Intermediate Representation for Automatic Text Summarization H.G. Silber and K.F. McCoy University of Delaware.
A Cross-Lingual ILP Solution to Zero Anaphora Resolution Ryu Iida & Massimo Poesio (ACL-HLT 2011)
Opinion Holders in Opinion Text from Online Newspapers Youngho Kim, Yuchul Jung and Sung-Hyon Myaeng Reporter: Chia-Ying Lee Advisor: Prof. Hsin-Hsi Chen.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
Ground Truth Free Evaluation of Segment Based Maps Rolf Lakaemper Temple University, Philadelphia,PA,USA.
A Semantic Approach to IE Pattern Induction Mark Stevenson and Mark A. Greenwood Natural Language Processing Group University of Sheffield, UK.
COLING 2012 Extracting and Normalizing Entity-Actions from Users’ comments Swapna Gottipati, Jing Jiang School of Information Systems, Singapore Management.
Wikipedia as Sense Inventory to Improve Diversity in Web Search Results Celina SantamariaJulio GonzaloJavier Artiles nlp.uned.es UNED,c/Juan del Rosal,
Inference Protocols for Coreference Resolution Kai-Wei Chang, Rajhans Samdani, Alla Rozovskaya, Nick Rizzolo, Mark Sammons, and Dan Roth This research.
Evaluation issues in anaphora resolution and beyond Ruslan Mitkov University of Wolverhampton Faro, 27 June 2002.
Support Vector Machines and Kernel Methods for Co-Reference Resolution 2007 Summer Workshop on Human Language Technology Center for Language and Speech.
Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.
FILTERED RANKING FOR BOOTSTRAPPING IN EVENT EXTRACTION Shasha Liao Ralph York University.
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
ICE: a GUI for training extraction engines CSCI-GA.2590 Ralph Grishman NYU.
Cold-Start KBP Something from Nothing Sean Monahan, Dean Carpenter Language Computer.
Solving Hard Coreference Problems Haoruo Peng, Daniel Khashabi and Dan Roth Problem Description  Problems with Existing Coref Systems Rely heavily on.
ANAPHORA RESOLUTION SYSTEM FOR NATURAL LANGUAGE REQUIREMENTS DOCUMENT IN KOREAN 課程 : 自然語言與應用 課程老師 : 顏國郎 報告者 : 鄭冠瑀.
Question Answering Passage Retrieval Using Dependency Relations (SIGIR 2005) (National University of Singapore) Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan,
Mastering the Pipeline CSCI-GA.2590 Ralph Grishman NYU.
gesture features for coreference
A Deep Memory Network for Chinese Zero Pronoun Resolution
Maximum Entropy Models and Feature Engineering CSCI-GA.2591
Simone Paolo Ponzetto University of Heidelberg Massimo Poesio
Preliminaries CSCI-GA.2591
Relation Extraction CSCI-GA.2591
Entity- & Topic-Based Information Ordering
(Entity and) Event Extraction CSCI-GA.2591
Dr. Clincy Professor of CS
Training and Evaluation CSCI-GA.2591
Background & Overview Proposed Model Experimental Results Future Work
Introduction to Information Extraction
Social Knowledge Mining
Lecture 9: Entity Resolution
Clustering Algorithms for Noun Phrase Coreference Resolution
Lei Sha, Jing Liu, Chin-Yew Lin, Sujian Li, Baobao Chang, Zhifang Sui
Dr. Clincy Professor of CS
An Interactive Approach to Collectively Resolving URI Coreference
Leverage Consensus Partition for Domain-Specific Entity Coreference
Attention for translation
Rachit Saluja 03/20/2019 Relation Extraction with Matrix Factorization and Universal Schemas Sebastian Riedel, Limin Yao, Andrew.
Topic: Semantic Text Mining
Presentation transcript:

NYU Coreference CSCI-GA.2591 Ralph Grishman

basically a clustering task: clustering mentions into entities

Types of Referring Expressions names nominals pronouns

Types of ‘Coreference’ Identity Predication ACE considers this part of coreference Bridging Anaphora Not included in ACE

Strategies mention-pair model entity-mention model entity-entity model for every pair of mentions, model determines coreferentiality then (to enforce transitivity) cluster mentions guided by these decisions entity-mention model single pass through document, building entities model chooses which entity to add mention to (if any) entity-entity model agglomerative clustering

Ordering clustering can be done in one pass in several passes (sieve) in dynamically-determined order Ragunathan et al. report gains of 1-4% in F1 score from multi-pass over single pass But multi-pass makes incremental processing more difficult

Hand-coded rules The bulk of the cases follow well-understood patterns: predicate complement apposition role modifier relative pronouns … So many systems use hand-coded rules or hybrid systems combining corpus-trained and hand-coded rules [Jet]

Hand-coded rules Sieve: exact extent match appositive | predicate nominative | role appositive | relative pronoun | acronym | demonym cluster head match with word inclusion and compatible modifiers cluster head match with word inclusion cluster head match with compatible modifiers relaxed cluster head match with word inclusion pronoun match [Ragunathan et al. 2010]

Hand-coded rules Sieve: P R F exact extent match 96 32 48 appositive | predicate nominative | role appositive | 95 44 60 relative pronoun | acronym | demonym cluster head match with word inclusion and 92 51 66 compatible modifiers cluster head match with word inclusion 92 52 66 cluster head match with compatible modifiers 91 53 67 relaxed cluster head match with word inclusion 90 54 67 pronoun match 84 74 79 nominal coref helps little

Anaphoricity Should we have a separate model for anaphoricity?

Role of deep learning Systems do worst in resolving nominal anaphors systems typically extract features of the anaphor and candidate antecedent and then use a log-linear model to capture compatibility for example, using WordNet lexical relations Deep learning systems try to do this more directly: building a large distributed representation of the mentions and the entities (based on the word embeddings of the words in and words in the immediate context of the anaphor and the candidate antecedent) and then learning a ranking among entity pairs [Clark and Manning 2016]

Benchmarks Most common evaluation is SemEval 2011 Based on OntoNotes corpus Did not mark singletons Included event references

Evaluation Metric There is no consensus on an evaluation metric for coreference SemEval used an average of 3 scores MUC score B-cubed CEAF (not to mention the official ACE scorer)

MUC Scoring The first coreference scorer was developed for MUC-6; it is link-based The key S and the response R each define a set of equivalence classes Si and Ri To assess the recall of the response with respect to class Si, we ask how many links would have to be added to R to link all the mentions in S = p(S)-1 Recalli = |Si – p(Si)| / |Si| - 1 Recall = sum of Recalli Precision is computed by swapping S and R

MUC Scoring Example: Truth 1 --- 2 --- 3 --- 4 --- 5 6 ---7 8 ---9 --- A --- B --- C Response 1 --- 2 --- 3 --- 4 --- 5

MUC Scoring One shortcoming of MUC scoring is that you don’t get credit for correct singletons Also, the metric rates as equal some responses which are worse than others

MUC Scoring Example: Truth 1 --- 2 --- 3 --- 4 --- 5 6 ---7 8 ---9 --- A --- B --- C Response 1 --- 2 --- 3 --- 4 --- 5 With MUC scorer, this gets the same score

B-CUBED Scoring B-CUBED is a mention-based scorer designed to avoid the problem of the MUC scorer [Bagga and Baldwin 1998] Precisioni = # of correct mentions in response equiv. class containing mentioni / # of mentions in response equiv class containing mentioni Recalli = # of correct mentions in response equiv class containing mentioni / # of mentions in key equiv class containing mentioni

CEAF Constrained Entity Alignment F-measure Based on a similarity metric between clusters (entities) Computes an optimal alignment between key and response using this metric Leftover clusters are not scored