Presentation is loading. Please wait.

Presentation is loading. Please wait.

Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F-78153 Le Chesnay, France Jason Baldridge.

Similar presentations


Presentation on theme: "Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F-78153 Le Chesnay, France Jason Baldridge."— Presentation transcript:

1 Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F-78153 Le Chesnay, France Jason Baldridge Department of Linguistics University of Texas at Austin In Proceedings of EMNLP-2008 1 Yi-Ting Huang 2010/12/24

2 Outline 1.Introduction 2.Anaphoric filter 3.Ranking 4.Specialized models 5.Experiment 6.Conclusion 7.Comments 2

3 1. Introduction Coreference resolution is the task of partitioning a set of entity mentions in a text, where each partition corresponds to some entity in an underlying discourse model. pair-wise classification (c1, m)=Y (c2, m)=N (c3, m)=N (c4, m)=Y (c5, m)=Y (c6, m)=N Combine entity1(c1, c4, c5, m) entity2(c2, c3) entity3(c6) 3 c1….c2……. …c3……c4.. ……c5…….. c6………m..

4 Research problem 1.Numerous approaches to anaphora and coreference resolution reduce these tasks to a binary classification task, whereby pairs of mentions are classified as coreferential or not. – these approaches make very strong independence assumptions. 2.Only one pair-wise classification to identify any type of noun phrases. 4 pair-wise classification (c1, m)=Y (c2, m)=N (c3, m)=N (c4, m)=Y (c5, m)=Y (c6, m)=N Combine entity1(c1, c4, c5, m) entity2(c2, c3) entity3(c6) c1….c2……. …c3……c4.. ……c5…….. c6………m..

5 However… (1/2) For many authors, the relation takes the form of a continuum and is often represented in the form of a referential hierarchy, such as: The higher up, the more accessible (or salient) the entity is. This type of hierarchy is validated by corpus studies of the distribution of different types of expressions. 5 M. Ariel. 1988. Referring and accessibility. Journal of Linguistics, pages 65–87.

6 However… (2/2) At the extremes are pronouns (these forms typically require a previous mention in the local context) and proper names (these forms are often used without previous mentions of the entity). For instance, pronouns find their antecedents very locally (in a window of 1-2 sentences), while proper names predominantly find theirs at longer distances (Ariel, 1988). 6

7 Research purpose The use of rankers instead of classifiers The use of linguistically motivated, specialized models for different types of mentions. – third-person pronouns, – speech pronouns (i.e., first and second person pronouns), – proper names, – definite descriptions – other types of nominals (e.g., anaphoric uses of indefinite, quantified, and bare noun phrases). 7

8 Procedure discourse status classifier Step 1 anaphoric filter m a part of mention a new entity Step 2 3 rd person pronouns Step 3 specialized models and ranker ((c1, c4, c5), m) c1….c2……. …c3……c4.. ……c5…….. c6………m.. ((c2, c3), m) ((c6), m) m proper namesothersSpeech pronounsdefinite descriptions identify m’s type Step 2 identify type 8 (c6, m)

9 Term definition referential expressions  mention a discourse entity  entity 9 c1….c2……. …c3……c4.. ……c5…….. c6………m.. Combine entity1(c1, c4, c5, m) entity2(c2, c3) entity3(c6)

10 Procedure discourse status classifier Step 1 anaphoric filter m a part of mention a new entity Step 2 3 rd person pronouns Step 3 specialized models and ranker ((c1, c4, c5), m) c1….c2……. …c3……c4.. ……c5…….. c6………m.. ((c2, c3), m) ((c6), m) m proper names others Speech pronouns definite descriptions identify m’s type Step 2 identify type 10 (c6, m)

11 2. anaphoric filter (1/2) Not all referential expressions in a given document are anaphors: some expressions introduce a discourse entity, rather than accessing an existing one. It is more troublesome for rankers, which always pick an antecedent from the candidate set. A natural solution is to use a model that specifically predicts the discourse status (discourse-new vs. discourse-old) of each expression: only expressions that are classified as “discourse-old” by this model are considered by rankers. 11

12 2. anaphoric filter (2/2) We use a similar discourse status classifier to Ng and Cardie’s (Ng and Cardie, 2002a) as a filter on mentions for our rankers. We rely on three main types of information sources: – the form of mention (e.g., type of linguistic expression, number of tokens), – positional features in the text, – comparisons of the given mention to the mentions that precede it in the text. Evaluated on the ACE datasets, the model achieves an overall accuracy score of 80.8%, compared to a baseline of 59.7% when predicting the majority class (“discourse-old”). 12 V. Ng and C. Cardie. 2002a. Identifying anaphoric and non-anaphoric noun phrases to improve coreference resolution. In Proceedings of COLING 2002.

13 Procedure discourse status classifier Step 1 anaphoric filter m a part of mention a new entity Step 2 c1….c2……. …c3……c4.. ……c5…….. c6………m.. 3 rd person pronouns Step 3 specialized models and ranker ((c1, c4, c5), m) ((c2, c3), m) ((c6), m) m proper namesothersSpeech pronounsdefinite descriptions identify m’s type Step 2 identify type 13 (c6, m)

14 3. Ranking (1/3) 14 c1….c2……. …c3……c4.. ……c5…….. c6………m..

15 3. Ranking (2/3) For the training of the different ranking models, we use the following procedure. For each model, instances are created by pairing each anaphor with a set of candidates which contains: – true antecedent – a set of non-antecedents 15

16 3. Ranking (3/3) True antecedent: The selection of the true antecedent varies depending on the model we are training: – for pronominal forms, the antecedent is selected as the closest preceding mention in the chain; – for non-pronominal forms, we used the closest preceding non-pronominal mention in the chain as the antecedent. Non-antecedent set: we collect all the non-antecedents that appear in a window of two sentences around the antecedent. 16

17 3. Specialized models (1/2) Our second strategy is to use different, specialized models for different referential expressions. We use separate models for the following types: 1.third person pronouns, 2.speech pronouns, 3.proper names, 4.definite descriptions, and 5.others. 17

18 3. Specialized models (2/2) Note that our split of referential types only partially cover the referential hierarchies of Ariel (1988). Thus, there is no separate model for demonstrative noun phrases and pronouns: these are very rare in the corpus we used (i.e., the ACE corpus). 18

19 3.1 features (1/6) The total number of anaphors (i.e., of mentions that are not chain heads) in the data is 19322 and 4599 for training and testing, respectively. 22-24% 11-13% 33-40% 16-17% 19

20 3.1 features (2/6) Our feature extraction relies on limited linguistic processing: we only made use of a sentence detector, a tokenizer, a POS tagger (as provided by the OpenNLP Toolkit) and the WordNet database. Since we did not use parser, lexical heads for the NP mentions were computed using simple heuristics relying solely on POS sequences. 20

21 3.1 features (3/6) 21

22 3.1 features (4/6) Linguistic form: Distance: Morph syntactic Agreement Semantic compatibility 22 we collected the synonym set (or synset) as well as the synset of their direct hypernyms associated with each mention. we used the synset associated with the first sense associated with the mention’s head word. we collected the synonym set (or synset) as well as the synset of their direct hypernyms associated with each mention. we used the synset associated with the first sense associated with the mention’s head word.

23 3.1 features (5/6) Context: – these features can be seen as approximations of the grammatical roles, as indicators of the salience of the potential candidate. the part of speech tags surrounding the candidate, indicates whether the potential antecedent is the first mention in a sentence (approximating subject-hood), indicating whether the candidate is embedded inside another mention. 23

24 3.1 features (6/6) String similarity: for PN, Def-NP, Other Acronym: for PN, Def-NP Apposition: for PN 24

25 5.1 Experiment 1 Purpose: their ability to select a correct antecedent for each anaphor. Accuracy: the ratio of correctly resolved anaphors. 25

26 5.2 Experiment 2 Metric: MUC, B 3, CEAF DataSet: ACE02 26 Ng and Cardie (2002b) V. Ng and C. Cardie. 2002b. Improving machine learning approaches to coreference resolution. In Proceedings of ACL 2002, pages 104–111.

27 5.2 error analysis missed anaphors – i.e., an anaphoric mention that fails to be linked to a previous mention spurious anaphors – i.e., an non-anaphoric mention that is linked to a previous mention invalid resolutions – a true anaphor that is linked to a incorrect antecedent 27

28 5.3 Experiment 3 RANK+DS-ORACLE+SP – uses the specialized rankers in combination with a perfect discourse status classifier. LINK-ORACLE – uses the discourse status classifier with a perfect coreference resolver. 28

29 6. Conclusion We present and evaluate two straight-forward tactics for improving coreference resolution: – ranking models, and – separate, specialized models for different types of referring expressions. This simple pipeline architecture produces significant improvements over various implementations of the standard, classifier- based coreference system. 29

30 7. Comments Advantage: – The proposed method by ranker and specialized model, and the result of 3 rd pronouns, proper names and link-oracle shows good performance. 30

31 Thank you for your attention & Merry Christmas! 31


Download ppt "Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F-78153 Le Chesnay, France Jason Baldridge."

Similar presentations


Ads by Google