Question Answering for Machine Reading Evaluation Evaluation Campaign at CLEF 2011 Anselmo Peñas (UNED, Spain) Eduard Hovy (USC-ISI, USA) Pamela Forner.

Slides:



Advertisements
Similar presentations
OMV Ontology Metadata Vocabulary April 10, 2008 Peter Haase.
Advertisements

Text Analysis Conference Knowledge Base Population 2013 Hoa Trang Dang National Institute of Standards and Technology Sponsored by:
ResPubliQA 2010: QA on European Legislation Anselmo Peñas, UNED, Spain Pamela Forner, CELCT, Italy Richard Sutcliffe, U. Limerick, Ireland Alvaro Rodrigo,
1 CLEF 2012, Rome QA4MRE, Question Answering for Machine Reading Evaluation Anselmo Peñas (UNED, Spain) Eduard Hovy (USC-ISI, USA) Pamela Forner (CELCT,
ISCOL 2011 – Bar Ilan University /151 A Probabilistic Model for Lexical Entailment Eyal Shnarch, Jacob Goldberger, Ido Dagan Bar Ilan University.
1 CLEF 2011, Amsterdam QA4MRE, Question Answering for Machine Reading Evaluation Question Answering Track Overview Main Task Anselmo Peñas Eduard Hovy.
Unsupervised Acquisition of Axioms to Paraphrase Noun Compounds and Genitives CICLING 2012, New Delhi Anselmo Peñas NLP & IR Group, UNED, Spain Ekaterina.
CLEF 2008 Multilingual Question Answering Track UNED Anselmo Peñas Valentín Sama Álvaro Rodrigo CELCT Danilo Giampiccolo Pamela Forner.
Textual Entailment Using Univariate Density Model and Maximizing Discriminant Function “Third Recognizing Textual Entailment Challenge 2007 Submission”
XML Document Mining Challenge Bridging the gap between Information Retrieval and Machine Learning Ludovic DENOYER – University of Paris 6.
3rd Answer Validation Exercise ( AVE 2008) QA subtrack at Cross-Language Evaluation Forum 2008 UNED Anselmo Peñas Álvaro Rodrigo Felisa Verdejo Thanks.
Open Information Extraction From The Web Rani Qumsiyeh.
Web Logs and Question Answering Richard Sutcliffe 1, Udo Kruschwitz 2, Thomas Mandl University of Limerick, Ireland 2 - University of Essex, UK 3.
CLEF 2007 Multilingual Question Answering Track Danilo Giampiccolo, CELCT Anselmo Peñas, UNED.
Enhance legal retrieval applications with an automatically induced knowledge base Ka Kan Lo.
Software Development Proposal Ho Chi Minh Rareblood Club Website Prepared by: BAKCO Date: 26-Oct-2009.
Semantic Web Technologies Lecture # 2 Faculty of Computer Science, IBA.
Spanish Question Answering Evaluation Anselmo Peñas, Felisa Verdejo and Jesús Herrera UNED NLP Group Distance Learning University of Spain CICLing 2004,
Overview of the Fourth Recognising Textual Entailment Challenge NIST-Nov. 17, 2008TAC Danilo Giampiccolo (coordinator, CELCT) Hoa Trang Dan (NIST)
Challenges in Information Retrieval and Language Modeling Michael Shepherd Dalhousie University Halifax, NS Canada.
Answer Validation Exercise - AVE QA subtrack at Cross-Language Evaluation Forum 2007 UNED (coord.) Anselmo Peñas Álvaro Rodrigo Valentín Sama Felisa Verdejo.
Russian Information Retrieval Evaluation Seminar (ROMIP) Igor Nekrestyanov, Pavel Braslavski CLEF 2010.
The Second PASCAL Recognising Textual Entailment Challenge Roy Bar-Haim, Ido Dagan, Bill Dolan, Lisa Ferro, Danilo Giampicollo, Bernardo Magnini, Idan.
“How much context do you need?” An experiment about context size in Interactive Cross-language Question Answering B. Navarro, L. Moreno-Monteagudo, E.
CLEF – Cross Language Evaluation Forum Question Answering at CLEF 2003 ( The Multiple Language Question Answering Track at CLEF 2003.
Prepared by Opinion Dynamics Corporation May 2004.
OWL Capturing Semantic Information using a Standard Web Ontology Language Aditya Kalyanpur Jennifer Jay Banerjee James Hendler Presented By Rami Al-Ghanmi.
RTE Planning Session Luisa Bentivogli, Peter Clark, Ido Dagan, Hoa Trang Dang, Danilo Giampiccolo.
Answer Validation Exercise - AVE QA subtrack at Cross-Language Evaluation Forum UNED (coord.) Anselmo Peñas Álvaro Rodrigo Valentín Sama Felisa Verdejo.
Cross-Language Evaluation Forum (CLEF) IST Expected Kick-off Date: August 2001 Carol Peters IEI-CNR, Pisa, Italy Carol Peters: blabla Carol.
EU Project proposal. Andrei S. Lopatenko 1 EU Project Proposal CERIF-SW Andrei S. Lopatenko Vienna University of Technology
Evaluating Multilingual Question Answering Systems at CLEF Pamela Forner 1, Danilo Giampiccolo 1, Bernardo Magnini 2, Anselmo Peñas 3, Álvaro Rodrigo 3,
Recognizing textual entailment: Rational, evaluation and approaches Source:Natural Language Engineering 15 (4) Author:Ido Dagan, Bill Dolan, Bernardo Magnini.
BAA - Big Mechanism using SIRA Technology Chuck Rehberg CTO at Trigent Software and Chief Scientist at Semantic Insights™
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
ICT TOOLS AND SOCIETY INVOLVEMENT AMONG THE EUPAN NETWORK HIGHLIGHTS FROM THE SURVEY RESULTS TANYA CHETCUTI AND MARCO FICHERA - WORKSHOP EUROPEAN COMMISSION.
For Monday Read chapter 24, sections 1-3 Homework: –Chapter 23, exercise 8.
For Monday Read chapter 26 Last Homework –Chapter 23, exercise 7.
OWL Representing Information Using the Web Ontology Language.
Trustworthy Semantic Webs Dr. Bhavani Thuraisingham The University of Texas at Dallas Lecture #4 Vision for Semantic Web.
Semantic web Bootstrapping & Annotation Hassan Sayyadi Semantic web research laboratory Computer department Sharif university of.
QA Pilot Task at CLEF 2004 Jesús Herrera Anselmo Peñas Felisa Verdejo UNED NLP Group Cross-Language Evaluation Forum Bath, UK - September 2004.
Evaluating Answer Validation in multi- stream Question Answering Álvaro Rodrigo, Anselmo Peñas, Felisa Verdejo UNED NLP & IR group nlp.uned.es The Second.
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
LREC Marrakech, May 29, 2008 Question Answering on Speech Transcriptions: the QAST evaluation in CLEF L. Lamel 1, S. Rosset 1, C. Ayache 2, D. Mostefa.
What’s happening in iCLEF? (the iCLEF Flickr Challenge) Julio Gonzalo (UNED), Paul Clough (U. Sheffield), Jussi Karlgren (SICS), Javier Artiles (UNED),
Toward an Open Source Textual Entailment Platform (Excitement Project) Bernardo Magnini (on behalf of the Excitement consortium) 1 STS workshop, NYC March.
Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.
Towards Entailment Based Question Answering: ITC-irst at Clef 2006 Milen Kouylekov, Matteo Negri, Bernardo Magnini & Bonaventura Coppola ITC-irst, Centro.
Relevance Models and Answer Granularity for Question Answering W. Bruce Croft and James Allan CIIR University of Massachusetts, Amherst.
Slide 1 Eurostat Unit B3 – Statistical Information Technology XBRL pilot Task Force 27 September 2006 Status of XDIS/XBRL pilot project Giuseppe Sindoni.
For Monday Read chapter 26 Homework: –Chapter 23, exercises 8 and 9.
Textual Analysis Introduction. What is Textual Analysis? Textual Analysis, as the name suggests, involves the Analysis of a literary Text. It is very.
Recognising Textual Entailment Johan Bos School of Informatics University of Edinburgh Scotland,UK.
WP1: Plan for the remainder (1) Ontology –Finalise ontology and lexicons for the 2 nd domain (RTV) Changes agreed in Heraklion –Improvement to existing.
11 Thoughts on STS regarding Machine Reading Ralph Weischedel 12 March 2012.
Trends in NL Analysis Jim Critz University of New York in Prague EurOpen.CZ 12 December 2008.
WP4 Models and Contents Quality Assessment
Sentiment analysis algorithms and applications: A survey
Guangbing Yang Presentation for Xerox Docushare Symposium in 2011
Presentation 王睿.
Text-to-Text Generation
What is the Entrance Exams Task
AP Biology Class Bonneville High School
UNED Anselmo Peñas Álvaro Rodrigo Felisa Verdejo Thanks to…
Information Retrieval
Topic: Semantic Text Mining
Machine Reading.
CLEF 2008 Multilingual Question Answering Track
Presentation transcript:

Question Answering for Machine Reading Evaluation Evaluation Campaign at CLEF 2011 Anselmo Peñas (UNED, Spain) Eduard Hovy (USC-ISI, USA) Pamela Forner (CELCT, Italy) Richard Sutcliffe (U. Limerick, Ireland) Álvaro Rodrigo (UNED, Spain)

UNED nlp.uned.es Knowledge-Understanding dependence We “understand” because we “know” Capture ‘knowledge’ expressed in texts ‘Understand’ language

UNED nlp.uned.es Control the variable of knowledge The ability of making inferences about texts is correlated to the amount of knowledge considered This variable has to be taken into account during evaluation Otherwise is very difficult to compare methods How to control the variable of knowledge in a reading task?

UNED nlp.uned.es Question Answering Restricted-domain QA systems 1. On large knowledge bases Structured QA, not aiming for language understanding 2. On a domain specific collection Information Extraction rules Open domain QA systems 1. On open domain collections Based on retrieval and redundancy Very limited inference What’s next in QA?

UNED nlp.uned.es Recognizing Textual Entailment Test: Text (evidence) – Hypothesis pair Source of knowledge: Free Difficult to evaluate if best systems have better methods or better knowledge or both Cheap evaluation Reusable 100% Same framework for any level of complexity What´s next in RTE? Control the variable of knowledge

UNED nlp.uned.es Proposal: QA4MRE Question Answering for Machine Reading Evaluation (QA4MRE) New task of QA Track at CLEF 2011 General Goal Measure progress in two reading abilities Capture knowledge from text collections Answer questions about a single text

UNED nlp.uned.es Requirements Don’t fix the representation formalism Semantic representation beyond sentence level is part of the research agenda Don't build systems tuned for specific domains But general technologies, able to self-adapt to new contexts or topics Evaluate reading abilities Knowledge acquisition Answer questions about a single document Control the role of knowledge

UNED nlp.uned.es Sources of knowledge Text Collection Big and diverse enough to acquire required knowledge Impossible for all possible topics Define a scalable strategy: topic by topic Several topics Narrow enough to limit knowledge needed (e.g. Petroleum industry, European Football League, Disarmament of the Irish Republican Army, etc.) Reference collection per topic (10,000-50,000 docs.) Documents defining concepts about the topic (e.g. wikipedia) News about the topic Web pages, blogs, opinions

UNED nlp.uned.es Reading test Text Coal seam gas drilling in Australia's Surat Basin has been halted by flooding. Australia's Easternwell, being acquired by Transfield Services, has ceased drilling because of the flooding. The company is drilling coal seam gas wells for Australia's Santos Ltd. Santos said the impact was minimal. Multiple choice test According to the text… What company owns wells in Surat Basin? a)Australia b)Coal seam gas wells c)Easternwell d)Transfield Services e)Santos Ltd. f)Ausam Energy Corporation g)Queensland h)Chinchilla

UNED nlp.uned.es Knowledge gaps Company A Company BWell C fordrills Owns | P Acquire this knowledge from the reference collection Queensland Australia Surat Basin is part of

UNED nlp.uned.es Runs Type I No external sources of knowledge Only the given reference collection Type II With external sources Specify which ones.

UNED nlp.uned.es Schedule Guidelines and Samples of tests1 st February Release of topics and reference corpora1 st April Test set release1 st June Run submissions15 th June Results to the participants1 st July Submission of Notebook PapersAugust - September Web site:

UNED nlp.uned.es Program Committee Ken Barker, University of Texas at Austin, US Johan Bos, Rijksuniversiteit Groningen, Netherlands Peter Clark, Vulcan Inc., US Ido Dagan, Bar-Ilan University, Israel Bernardo Magnini, Fondazione Bruno Kessler, Italy Dan Moldovan, University of Texas at Dallas, US Emanuele Pianta, Fondazione Bruno Kessler, and CELCT, Italy John Prager, IBM, US Dan Tufis, RACAI, Romania Hoa Trang Dang, NIST, US

UNED nlp.uned.es Join the organization Working group is open to collaboration Development collections Add new languages Define types of questions Write down tests about a topic …