Assessing the Impact of Frame Semantics on Textual Entailment Authors: Aljoscha Burchardt, Marco Pennacchiotti, Stefan Thater, Manfred Pinkal Saarland.

Slides:



Advertisements
Similar presentations
COGEX at the Second RTE Marta Tatu, Brandon Iles, John Slavick, Adrian Novischi, Dan Moldovan Language Computer Corporation April 10 th, 2006.
Advertisements

COGEX at the Second RTE Marta Tatu, Brandon Iles, John Slavick, Adrian Novischi, Dan Moldovan Language Computer Corporation April 10 th, 2006.
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Recovering Human Body Configurations: Combining Segmentation and Recognition Greg Mori, Xiaofeng Ren, and Jitentendra Malik (UC Berkeley) Alexei A. Efros.
MEANT: semi-automatic metric for evaluating for MT evaluation via semantic frames an asembling of ACL11,IJCAI11,SSST11 Chi-kiu Lo & Dekai Wu Presented.
Overview of the Hindi-Urdu Treebank Fei Xia University of Washington 7/23/2011.
FATE: a FrameNet Annotated corpus for Textual Entailment Marco Pennacchiotti, Aljoscha Burchardt Computerlinguistik Saarland University, Germany LREC 2008,
Max-Margin Matching for Semantic Role Labeling David Vickrey James Connor Daphne Koller Stanford University.
Semantic Role Labeling Abdul-Lateef Yussiff
FATE: a FrameNet Annotated corpus for Textual Entailment Marco Pennacchiotti, Aljoscha Burchardt Computerlinguistik Saarland University, Germany LREC 2008,
Steven Schoonover.  What is VerbNet?  Levin Classification  In-depth look at VerbNet  Evolution of VerbNet  What is FrameNet?  Applications.
Logic in general Logics are formal languages for representing information such that conclusions can be drawn Syntax defines the sentences in the language.
Class Project Due at end of finals week Essentially anything you want, so long as its AI related and I approve Any programming language you want In pairs.
Logical Agents Chapter 7. Why Do We Need Logic? Problem-solving agents were very inflexible: hard code every possible state. Search is almost always exponential.
Logical Agents Chapter 7. Why Do We Need Logic? Problem-solving agents were very inflexible: hard code every possible state. Search is almost always exponential.
1 QA in Discussion Boards  Companies (e.g., Dell, IBM) use discussion boards as ways for customers to get answers to their questions  90% of 40 analyzed.
Language, Mind, and Brain by Ewa Dabrowska Chapter 2: Language processing: speed and flexibility.
Natural Language Query Interface Mostafa Karkache & Bryce Wenninger.
Semi-Automatic Learning of Transfer Rules for Machine Translation of Low-Density Languages Katharina Probst April 5, 2002.
Comments on Guillaume Pitel: “Using bilingual LSA for FrameNet annotation of French text from generic resources” Gerd Fliedner Computational Linguistics.
Agents that Reason Logically Logical agents have knowledge base, from which they draw conclusions TELL: provide new facts to agent ASK: decide on appropriate.
Meaning and Language Part 1.
Extracting Opinions, Opinion Holders, and Topics Expressed in Online News Media Text Soo-Min Kim and Eduard Hovy USC Information Sciences Institute 4676.
Outline P1EDA’s simple features currently implemented –And their ablation test Features we have reviewed from Literature –(Let’s briefly visit them) –Iftene’s.
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Syntax Lecture 8: Verb Types 1. Introduction We have seen: – The subject starts off close to the verb, but moves to specifier of IP – The verb starts.
BİL711 Natural Language Processing1 Statistical Parse Disambiguation Problem: –How do we disambiguate among a set of parses of a given sentence? –We want.
PropBank, VerbNet & SemLink Edward Loper. PropBank 1M words of WSJ annotated with predicate- argument structures for verbs. –The location & type of each.
The Impact of Grammar Enhancement on Semantic Resources Induction Luca Dini Giampaolo Mazzini
SALSA The Saarbrücken Lexical Semantics Annotation & Acquisition Project Aljoscha Burchardt, Katrin Erk, Anette Frank, Andrea Kowalski, Sebastian Pado,
Interpreting Dictionary Definitions Dan Tecuci May 2002.
1 Logical Agents CS 171/271 (Chapter 7) Some text and images in these slides were drawn from Russel & Norvig’s published material.
Class Project Due at end of finals week Essentially anything you want, so long as its AI related and I approve Any programming language you want In pairs.
Modelling Human Thematic Fit Judgments IGK Colloquium 3/2/2005 Ulrike Padó.
11 Chapter 19 Lexical Semantics. 2 Lexical Ambiguity Most words in natural languages have multiple possible meanings. –“pen” (noun) The dog is in the.
Computational Semantics Day 5: Inference Aljoscha.
Bootstrapping for Text Learning Tasks Ramya Nagarajan AIML Seminar March 6, 2001.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
Conditionals Sarah Morris. What is a conditional?  A conditional sentence is a sentence containing the word if.  Something will happen if a condition.
Combining Lexical Resources: Mapping Between PropBank and VerbNet Edward Loper,Szu-ting Yi, Martha Palmer September 2006.
Rules, Movement, Ambiguity
GermaNet-WS II A WordNet “Detour” to FrameNet Aljoscha Burchardt Katrin Erk Anette Frank* Saarland University, DFKI* Saarbrücken
1 Logical Agents CS 171/271 (Chapter 7) Some text and images in these slides were drawn from Russel & Norvig’s published material.
Supertagging CMSC Natural Language Processing January 31, 2006.
CPSC 422, Lecture 27Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27 Nov, 16, 2015.
CS 4705 Lecture 17 Semantic Analysis: Robust Semantics.
FILTERED RANKING FOR BOOTSTRAPPING IN EVENT EXTRACTION Shasha Liao Ralph York University.
1 Fine-grained and Coarse-grained Word Sense Disambiguation Jinying Chen, Hoa Trang Dang, Martha Palmer August 22, 2003.
SALSA-WS 09/05 Approximating Textual Entailment with LFG and FrameNet Frames Aljoscha Burchardt, Anette Frank Computational Linguistics Department Saarland.
BLUE (Boeing Language Understanding Engine) - A Quick Tutorial on How it Works Working Note Peter Clark Phil Harrison (Boeing Phantom Works)
1 UNIT-3 KNOWLEDGE REPRESENTATION. 2 Agents that reason logically(Logical agents) A Knowledge based Agent The Wumpus world environment Representation,
Learning Event Durations from Event Descriptions Feng Pan, Rutu Mulkar, Jerry R. Hobbs University of Southern California ACL ’ 06.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
Meaning and Language Part 1. Plan We will talk about two different types of meaning, corresponding to two different types of objects: –Lexical Semantics:
Lec. 10.  In this section we explain which constituents of a sentence are minimally required, and why. We first provide an informal discussion and then.
Relation Extraction (RE) via Supervised Classification See: Jurafsky & Martin SLP book, Chapter 22 Exploring Various Knowledge in Relation Extraction.
COSC 6336: Natural Language Processing
Approaches to Machine Translation
Coarse-grained Word Sense Disambiguation
CSC 594 Topics in AI – Natural Language Processing
Chapter 11: Artificial Intelligence
Session 7: Face Detection (cont.)
Two Discourse Driven Language Models for Semantics
Logical Agents Chapter 7.
Automatic Detection of Causal Relations for Question Answering
Approaches to Machine Translation
Logical Agents Chapter 7.
Structure of a Lexicon Debasri Chakrabarti 13-May-19.
VERB PHYSICS: Relative Physical Knowledge of Actions and Objects
Presentation transcript:

Assessing the Impact of Frame Semantics on Textual Entailment Authors: Aljoscha Burchardt, Marco Pennacchiotti, Stefan Thater, Manfred Pinkal Saarland Univ, Germany in Natural Language Engineering 1 (1) pp1-25 As (mis-)interpreted by Peter Clark

The Textual Entailment Task  Syntactically, the players have moved.  from a syntactic point of view, T and H differ  But semantically, the players are still the same  from a semantic point of view, T and H are the same  So, want to identify and match on semantic, not syntactic, level:  Need for “frame semantics”  (syntax) X drown Y → (semantics) cause* drown victim  (syntax) X drown in Y → (semantics) victim drown in cause T: A flood drowned 11 people. H: 11 people drowned in a flood. T: flood drown people H: people drown in flood * if X is inanimate (otherwise role is killer)

Frame Semantic Resources  PropBank:  thematic roles (arg0, arg1, …):  arg0 search arg1 for arg2 (“Mary searched the room for the ring”)  arg0 search for arg2 (“Fred searched for the ring”)  BUT roles are verb-specific (and names are overloaded)  arg0 seek arg1 (“Mary sought the ring”)  No guarantee arg0 means the same in different verbs  Note: thematic roles like “agent” are necessarily verb- specific:  Fred sold a car to John. John bought a car from Fred.  Thematic roles: Fred, John are both agents.  Case/semantic roles: Fred is the buyer, John is the seller.

Frame Semantic Resources  FrameNet:  Semantic roles are shared among verbs  several verbs map to the same Frame  Frames organized in a taxonomy  Roles organized in a taxonomy  Doesn’t contain subcategorization templates for semantic role labeling  causer kill victim  But does contain role-labeled examples, from which semantic role labeling algorithms can be learned

Example Frame in FrameNet

 So: it seems FrameNet should really help!

 Even more, FrameNet has (limited) inferential connections T: Wyniemko, now 54 and living in Rochester Hills, was arrested and tried in 1994 for a rape in Clinton Township. H: Wyniemko was accused of rape.

But, limited success in practice  PropBank used by several systems, including the RTE3 winner:  but unclear how much PropBank contributed  FrameNet used in SALSA (Burchardt and Frank)  Shalmaneser + Detour for Semantic Role Labeling (SRL)  (Detour boosts SRL when training examples are missing)  SALSA:  find matching semantic roles  see if the role fillers match  machine learning approach:  for set of known matching fillers  (i) compute features  (ii) learn which weighted sum of features implies match  But SALSA didn’t do significantly better than simple lexical overlap

Possible reasons for “failure”  Poor coverage of FrameNet  Decision of applicable Frame is poor  Semantic Role Labeling is poor  Role filler matching is poor  How to distinguish between these?  Create FATE, an annotated RTE corpus  Only annotated the “relevant” parts of the sentences

FATE  Annotated RTE2 corpus (400+ve, 400-ve exs)  Good interannotator agreement  ~2 months work to create  4488 frames, 9512 roles annotated in the corpus  includes 373 (8%) Unknown_Frame  1% Unknown_Role  → FrameNet coverage is good for this data!  Still, not always clear-cut:  Annotator: EXPORT; Shalmaneser: SENDING  SENDING is still plausible Cars exported by Japan increased

1. How do automatic and manual annotation compare? Does SALSA pick the right frame? When it picks the right frame, does assign the right roles? When it picks the right frame and role, does it get the right filler (i.e., the same head noun as the gold standard)   Fred sold the book on the shelf to Mary sellergoodsbuyer Commercial_Transaction

Results  If H is entailed by T, then we expect 1.The Frame for H to also be in T 2.The Frame’s roles used in H to also be in T 3.The role fillers in H to match those in T  These may also be true if H isn’t entailed by T  BUT: presumably with less probability

Results  If H is entailed by T, then we expect 1.The Frame for H to also be in T (more often) 2.The Frame’s roles used in H to also be in T (more often) 3.The role fillers in H to match those in T (more often)  These may also be true if H isn’t entailed by T  BUT: presumably with less probability  Also: compare with simple word overlap

Results  If H is entailed by T, then we expect 1.The Frame for H to also be in T (more often) Yes…. (Note: low difference here reflects that T and H typically talk about the same thing)

Results  If H is entailed by T, then we expect 1.The Frame for H to also be in T (more often) …but not much more than word overlap… (Not really surprising, as frames are picked based on words)

Results  If H is entailed by T, then we expect 1.The Frame for H to also be in T (more often) Also the hierarchy doesn’t help much here

Results  If H is entailed by T, then we expect 1.The Frame for H to also be in T (more often) 2.The Frame’s roles used in H to also be in T (more often) Again, low difference suggests that the roles talked about in T and H are usually the same For pairs which have a Frame in common between T and H:

Results  If H is entailed by T, then we expect 1.The Frame for H to also be in T (more often) 2.The Frame’s roles used in H to also be in T (more often) 3.The role fillers in H to match those in T (more often) 

Results  If H is entailed by T, then we expect 1.The Frame for H to also be in T (more often) 2.The Frame’s roles used in H to also be in T (more often) 3.The role fillers in H to match those in T (more often) T: An avalanche has struck a popular skiing resort in Australia, killing at least 11 people. H: Humans died in an avalanche. T: Virtual reality is used to train surgeons, pilots, astronauts, police officers, first-responders, and soldiers. H: Soldiers are trained using virtual reality. student Some difficult cases:

Results  Also, even if we had perfect frame, role, and filler matching, entailment does not always follow:  Negation:  Modality:

Conclusions 1.FrameNet’s coverage is good 2.Frame Semantic Analysis (frame/role/filler selection) is mediocre  3.Simple lexical overlap at the frame level don’t outperform simple lexical overlap at the syntactic level  4.Need better modeling:  wider context (negation, modalities)  role filler matching (semantic matching, e.g., WordNet)  more knowledge in FrameNet, e.g., implications  e.g., kill → die, arrest → accuse

(Extra slides)

The Textual Entailment Task: More complex example  Again, need to match semantic roles:  Again need for “frame semantics”  (syntax) X kill Y → (semantics) cause kill victim  (syntax) X died in Y → (semantics) protagonist died in cause  ALSO:  progagonist isa victim, Killing → Death T: An avalanche has struck a popular skiing resort in Australia, killing at least 11 people. H: Humans died in an avalanche. T: avalanche kill people H: human die in avalanche