Recognizing Textual Entailment Challenge PASCAL Suleiman BaniHani.

Slides:



Advertisements
Similar presentations
COGEX at the Second RTE Marta Tatu, Brandon Iles, John Slavick, Adrian Novischi, Dan Moldovan Language Computer Corporation April 10 th, 2006.
Advertisements

COGEX at the Second RTE Marta Tatu, Brandon Iles, John Slavick, Adrian Novischi, Dan Moldovan Language Computer Corporation April 10 th, 2006.
QA-LaSIE Components The question document and each candidate answer document pass through all nine components of the QA-LaSIE system in the order shown.
Proceedings of the Conference on Intelligent Text Processing and Computational Linguistics (CICLing-2007) Learning for Semantic Parsing Advisor: Hsin-His.
Baselines for Recognizing Textual Entailment Ling 541 Final Project Terrence Szymanski.
Background information Formal verification methods based on theorem proving techniques and model­checking –to prove the absence of errors (in the formal.
For Friday No reading Homework –Chapter 23, exercises 1, 13, 14, 19 –Not as bad as it sounds –Do them IN ORDER – do not read ahead here.
Robust Textual Inference via Graph Matching Aria Haghighi Andrew Ng Christopher Manning.
Research topics Semantic Web - Spring 2007 Computer Engineering Department Sharif University of Technology.
Normalized alignment of dependency trees for detecting textual entailment Erwin Marsi & Emiel Krahmer Tilburg University Wauter Bosma & Mariët Theune University.
Logic in general Logics are formal languages for representing information such that conclusions can be drawn Syntax defines the sentences in the language.
Gimme’ The Context: Context- driven Automatic Semantic Annotation with CPANKOW Philipp Cimiano et al.
Basi di dati distribuite Prof. M.T. PAZIENZA a.a
Automatic Classification of Semantic Relations between Facts and Opinions Koji Murakami, Eric Nichols, Junta Mizuno, Yotaro Watanabe, Hayato Goto, Megumi.
Introduction to CL Session 1: 7/08/2011. What is computational linguistics? Processing natural language text by computers  for practical applications.
UNED at PASCAL RTE-2 Challenge IR&NLP Group at UNED nlp.uned.es Jesús Herrera Anselmo Peñas Álvaro Rodrigo Felisa Verdejo.
Third Recognizing Textual Entailment Challenge Potential SNeRG Submission.
1 Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang, Assistant Professor Dept. of Computer Science & Information Engineering National Central.
Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang National Central University
Logical Agents Chapter 7 Feb 26, Knowledge and Reasoning Knowledge of action outcome enables problem solving –a reflex agent can only find way from.
Course Summary LING 575 Fei Xia 03/06/07. Outline Introduction to MT: 1 Major approaches –SMT: 3 –Transfer-based MT: 2 –Hybrid systems: 2 Other topics.
Article by: Feiyu Xu, Daniela Kurz, Jakub Piskorski, Sven Schmeier Article Summary by Mark Vickers.
Employing Two Question Answering Systems in TREC 2005 Harabagiu, Moldovan, et al 2005 Language Computer Corporation.
Overview of the Fourth Recognising Textual Entailment Challenge NIST-Nov. 17, 2008TAC Danilo Giampiccolo (coordinator, CELCT) Hoa Trang Dan (NIST)
Answer Validation Exercise - AVE QA subtrack at Cross-Language Evaluation Forum 2007 UNED (coord.) Anselmo Peñas Álvaro Rodrigo Valentín Sama Felisa Verdejo.
Knowledge and Tree-Edits in Learnable Entailment Proofs Asher Stern and Ido Dagan (earlier partial version by Roy Bar-Haim) Download at:
Copyright R. Weber Machine Learning, Data Mining ISYS370 Dr. R. Weber.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Knowledge and Tree-Edits in Learnable Entailment Proofs Asher Stern, Amnon Lotan, Shachar Mirkin, Eyal Shnarch, Lili Kotlerman, Jonathan Berant and Ido.
Chapter 10: Compilers and Language Translation Invitation to Computer Science, Java Version, Third Edition.
METEOR-Ranking & M-BLEU: Flexible Matching & Parameter Tuning for MT Evaluation Alon Lavie and Abhaya Agarwal Language Technologies Institute Carnegie.
Syntactic and semantic models and algorithms in Question Answering Alexander Solovyev Bauman Moscow Sate Technical University RCDL.
Scott Duvall, Brett South, Stéphane Meystre A Hands-on Introduction to Natural Language Processing in Healthcare Annotation as a Central Task for Development.
Machine Translation  Machine translation is of one of the earliest uses of AI  Two approaches:  Traditional approach using grammars, rewrite rules,
Recognizing textual entailment: Rational, evaluation and approaches Source:Natural Language Engineering 15 (4) Author:Ido Dagan, Bill Dolan, Bernardo Magnini.
1 Knowledge Representation. 2 Definitions Knowledge Base Knowledge Base A set of representations of facts about the world. A set of representations of.
1 Learning Sub-structures of Document Semantic Graphs for Document Summarization 1 Jure Leskovec, 1 Marko Grobelnik, 2 Natasa Milic-Frayling 1 Jozef Stefan.
Approaches to Machine Translation CSC 5930 Machine Translation Fall 2012 Dr. Tom Way.
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
1/21 Automatic Discovery of Intentions in Text and its Application to Question Answering (ACL 2005 Student Research Workshop )
For Monday Read chapter 24, sections 1-3 Homework: –Chapter 23, exercise 8.
For Monday Read chapter 26 Last Homework –Chapter 23, exercise 7.
Logical Agents Chapter 7. Outline Knowledge-based agents Logic in general Propositional (Boolean) logic Equivalence, validity, satisfiability.
Element Level Semantic Matching Pavel Shvaiko Meaning Coordination and Negotiation Workshop, ISWC 8 th November 2004, Hiroshima, Japan Paper by Fausto.
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
Collocations and Terminology Vasileios Hatzivassiloglou University of Texas at Dallas.
Automatic Question Answering  Introduction  Factoid Based Question Answering.
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
Evaluating Answer Validation in multi- stream Question Answering Álvaro Rodrigo, Anselmo Peñas, Felisa Verdejo UNED NLP & IR group nlp.uned.es The Second.
© Copyright 2008 STI INNSBRUCK Intelligent Systems Propositional Logic.
Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.
Improved Video Categorization from Text Metadata and User Comments ACM SIGIR 2011:Research and development in Information Retrieval - Katja Filippova -
Towards Entailment Based Question Answering: ITC-irst at Clef 2006 Milen Kouylekov, Matteo Negri, Bernardo Magnini & Bonaventura Coppola ITC-irst, Centro.
Generating Query Substitutions Alicia Wood. What is the problem to be solved?
SALSA-WS 09/05 Approximating Textual Entailment with LFG and FrameNet Frames Aljoscha Burchardt, Anette Frank Computational Linguistics Department Saarland.
Concepts and Realization of a Diagram Editor Generator Based on Hypergraph Transformation Author: Mark Minas Presenter: Song Gu.
Formal Verification. Background Information Formal verification methods based on theorem proving techniques and model­checking –To prove the absence of.
For Monday Read chapter 26 Homework: –Chapter 23, exercises 8 and 9.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
Recognising Textual Entailment Johan Bos School of Informatics University of Edinburgh Scotland,UK.
Logical Agents. Inference : Example 1 How many variables? 3 variables A,B,C How many models? 2 3 = 8 models.
Logical Agents. Outline Knowledge-based agents Logic in general - models and entailment Propositional (Boolean) logic Equivalence, validity, satisfiability.
Approaches to Machine Translation
Recognizing Partial Textual Entailment
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 20
Approaches to Machine Translation
CS246: Information Retrieval
ece 720 intelligent web: ontology and beyond
Information Retrieval
Presentation transcript:

Recognizing Textual Entailment Challenge PASCAL Suleiman BaniHani

Textual Entailment  Textual entailment recognition: is the task of deciding, given two text fragments, whether the meaning of one text is entailed (can be inferred) from another text.

Task Definition  Given pairs of small text snippets, referred to as Text- Hypothesis (T-H) pairs. Build a system that will decide for each T-H pair whether T indeed entails H or not. Results will be compared to the manual gold standard generated by annotators.  Example: T: Kurdistan Regional Government Prime Minister Dr. Barham Salih was unharmed after an assassination attempt. Prime minister targeted for assassination

Dataset Collection and Application Settings  The dataset of Text-Hypothesis pairs was collected by human annotators. It consists of seven subsets Information Retrieval (IR) Comparable Documents (CD) Reading Comprehension (RC) Question Answering (QA) Information Extraction (IE) Machine Translation (MT) Paraphrase Acquisition (PP)

Approaching textual entailment recognition  Solution approaches can be categorizes as. 1.Deep analysis or “understanding” –Using types of linguistic knowledge and resources to accurately recognize textual entailment Patterns of entailment (e.g. lexical relations, syntactic alternations) Processing technology (word co-occurrence statistics, thesaurus, parsing, etc.) 2.Shallow approach

Baseline  Given that half the pairs are FALSE, the simplest baseline is to label all pairs FALSE. This will achieve 50% accuracy.

Application of the BLEU (BiLingual Evaluation Understudy) algorithm  Shallow based on lexical level. It is based on calculating the percentage of n-grams for a given translation to the human standard one, a typical values for N are taken, i.e. 1, 2, 3, 4. It limits each n-gram appearance to a maximum frequency. The result of each n-gram is combined, and a penalty is added to short text.  Scored 54% for development set, and a 50% in the test set.  Good results in the CD, bad results in IE and IR.  Problem, does not recognize syntactical or semantics, such as synonyms and antonyms.

Syntactic similarities  Human annotators were asked to divide the data set to True by syntax False by syntax Not syntax Cannot decide  Then using a robust parser to establish the results.  A partial submission was provided. And humans were used for the test.

Tree edit distance  The text as well as the Hypothesis is transformed to a tree using a sentence splitter and a parser to create the syntactic representation.  A matching module, find the beast sequence of editing operations to obtain H from T.  Each editing operation (deletion, insertion and substitution) is given a relative score.  Finally the total score is evaluated, if it exceeds a certain limit them the pair is labeled as true.  High accuracy for CD but 55% overall accuracy.  Should be enriched by using resources as WordNet and other libraries.

Dependency Analysis and WordNet  A dependency parser is used to normalized data in appropriate tree representation.  Then a lexical entailment module is used, where the sub branches of T an H can be entailed from the other using Synonymy and similarity Hyponymy and WordNet entailment, i.e. death entail kill. Multiwords, i.e. melanoma entails skin-cancer. Negation and antonymy, where negation is propagated through tree leaves.  A matching between dependency trees using a matching algorithm searching for matching branches between T and H.  Results show high score in CD and a between 42 to 55 % in other fields

Syntactic Graph Distance: a rule based and a SVM based approach  Use a graph distance theory, where a graph is used to represent the H and T pair.  Use similarity measures to determine entailment T semantically subsumes H, e.g. H: [The cat eats the mouse] and T: [the cat devours the mouse], eat generalizes devour). T syntactically subsumes H, e.g., H: [The cat eats the mouse] and T: [the cat eats the mouse in the garden], T contains a specializing prepositional phrase). T directly implies H (e.g., H: [The cat killed the mouse], T: [the cat devours the mouse]).

Cont.  A rule based system realize the following Node similarity Syntactic similarity Semantic similarity  Applying a machine learning technique to evaluate the parameters and make the final decision  Results high for CD.76 and for others

hierarchical knowledge representation  A hierarchical logic passed representation o the T H pairs, where a description logic inspired language is used, extended feature description login (EFDL) which is similar to concept graph.  Nodes in the graph represent words or phrases.  Manually generated rewriting rules are used for semantic and syntactic representations.  A sentence in the text can have different alternatives  The evaluation is based if any of the sentence representations can infer H.  Results in the system set 64.8 while in the test 56.1, high CD lowest QA 50%

Logic like formula representation  A parser is used to transfer the pair T and H to graph, of logical phrases, where the nodes are the words and the links are the relations.  A matching score is given for each pair of terms.  The theorem proof is used to find the proof with the lowest coast.  The final cost is evaluated is it is less than a threshold, then the entailment is proved.  High results in the CD 79%, Lowest with MT 47% average 55%.

Atomic Propositions  Find entailment relation by comparing the atomic proposition contained in the T and H.  The comparison of the atomic propositions is done using a deduction system OTTER.  The atomic propositions are extracted from the text using a parser.  WordNet is used for word relations.  A semantic analyzer is used to transform the output of the parser to first order logic.  Low accuracy.5188 especially for QA 47%.

Combining shallow over lapping technique with deep theorem proving  In the shallow stage a simple frequency test of over lapping words is used.  In the deep stage CCG–parser is used to generate DRS, discourse representation theory. Which is transformed to first order logic.  Vampire theorem prover and Paradox where used for entailment proof.  A knowledge base was used to validate results with real world. WordNet Geographical knowledge from CIA Generic axioms for, for instance, the semantics of possessives, active-passives, and locations.  The combined system has accuracy of.562 while the shallow approach has an accuracy of 0.55.

Applying COGEX logic prover  First use parser to convert into logic.  Then use COGEX, which is a modified version of OTTER.  The prover requires a set of clauses called the “set of support” which is used to initiate the search for inferences.  The set of support is loaded with the negated form of the hypothesis as well as the predicates that make up the text passage.  Another list is required called the usable list, contains clauses used by OTTER to generate inferences.  The usable list consists of all the axioms that have been generated either automatically or by hand. World Knowledge Axioms (Manually) NLP Axioms(SS and SM) WordNet Lexical Chains  Overall accuracy.551, a lot of errors in the parsing stage.

Comparing task accuracy CD – Comparable DocumentsIE – Information Extraction QA – Question AnsweringIR – Information Retrieval MT – Machine TranslationRC – Reading Comprehension PP – Paraphrasing

Rsullt comparison

Future work  Search for a candidate parser to transform NL to first order logic.  Use the largest set of KB to caputre similarity.  Search for a robust theorem prover.

References  The first PASCAL Recognising Textual Entailment Challenge (RTE I)  Ido Dagan, Oren Glickman and Bernardo Magnini. The PASCAL Recognising Textual Entailment Challenge. In Proceedings of the PASCAL Challenges Workshop on Recognising Textual Entailment,  ex.html ex.html