AnswerFinder Question Answering from your Desktop Mark A. Greenwood Natural Language Processing Group Department of Computer Science University of Sheffield,

Slides:



Advertisements
Similar presentations
A complete citation, notecard, and outlining tool
Advertisements

QA-LaSIE Components The question document and each candidate answer document pass through all nine components of the QA-LaSIE system in the order shown.
WWW 2014 Seoul, April 8 th SNOW 2014 Data Challenge Two-level message clustering for topic detection in Twitter Georgios Petkos, Symeon Papadopoulos, Yiannis.
Question Answering Question Answering Available from: Mark A. Greenwood MEng.
Natural Language Processing Group Department of Computer Science University of Sheffield, UK IR4QA: An Unhappy Marriage Mark A. Greenwood.
1 Question Answering in Biomedicine Student: Andreea Tutos Id: Supervisor: Diego Molla.
NaLIX: A Generic Natural Language Search Environment for XML Data Presented by: Erik Mathisen 02/12/2008.
Chapter 2: Algorithm Discovery and Design
The Informative Role of WordNet in Open-Domain Question Answering Marius Paşca and Sanda M. Harabagiu (NAACL 2001) Presented by Shauna Eggers CS 620 February.
Employing Two Question Answering Systems in TREC 2005 Harabagiu, Moldovan, et al 2005 Language Computer Corporation.
Chapter 2: Algorithm Discovery and Design
Slide 3.1 Saunders, Lewis and Thornhill, Research Methods for Business Students, 5 th Edition, © Mark Saunders, Philip Lewis and Adrian Thornhill 2009.
Guidelines for Examination Candidates Raymond Hickey English Linguistics University of Duisburg and Essen (August 2015)
© Curriculum Foundation1 Section 2 The nature of the assessment task Section 2 The nature of the assessment task There are three key questions: What are.
A Light-weight Approach to Coreference Resolution for Named Entities in Text Marin Dimitrov Ontotext Lab, Sirma AI Kalina Bontcheva, Hamish Cunningham,
A Pattern Based Approach to Answering Factoid, List and Definition Questions Mark A. Greenwood and Horacio Saggion Natural Language Processing Group Department.
Challenges in Information Retrieval and Language Modeling Michael Shepherd Dalhousie University Halifax, NS Canada.
Systems Analysis – Analyzing Requirements.  Analyzing requirement stage identifies user information needs and new systems requirements  IS dev team.
How to Write a Literature Review
Computational Methods to Vocalize Arabic Texts H. Safadi*, O. Al Dakkak** & N. Ghneim**
Building a Simple Question Answering System Mark A. Greenwood Natural Language Processing Group Department of Computer Science University of Sheffield,
Probabilistic Model for Definitional Question Answering Kyoung-Soo Han, Young-In Song, and Hae-Chang Rim Korea University SIGIR 2006.
Chapter 2: Algorithm Discovery and Design Invitation to Computer Science, C++ Version, Third Edition.
The CoNLL-2013 Shared Task on Grammatical Error Correction Hwee Tou Ng, Yuanbin Wu, and Christian Hadiwinoto 1 Siew.
AnswerBus Question Answering System Zhiping Zheng School of Information, University of Michigan HLT 2002.
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
A Data Driven Approach to Query Expansion in Question Answering Leon Derczynski, Robert Gaizauskas, Mark Greenwood and Jun Wang Natural Language Processing.
A Markov Random Field Model for Term Dependencies Donald Metzler W. Bruce Croft Present by Chia-Hao Lee.
Modern Information Retrieval: A Brief Overview By Amit Singhal Ranjan Dash.
A Probabilistic Graphical Model for Joint Answer Ranking in Question Answering Jeongwoo Ko, Luo Si, Eric Nyberg (SIGIR ’ 07) Speaker: Cho, Chin Wei Advisor:
Search - on the Web and Locally Related directly to Web Search Engines: Part 1 and Part 2. IEEE Computer. June & August 2006.
Natural Language Based Reformulation Resource and Web Exploitation for Question Answering Ulf Hermjakob, Abdessamad Echihabi, Daniel Marcu University of.
Péter Schönhofen – Ad Hoc Hungarian → English – CLEF Workshop 20 Sep 2007 Performing Cross-Language Retrieval with Wikipedia Participation report for Ad.
A Language Independent Method for Question Classification COLING 2004.
XP New Perspectives on The Internet, Sixth Edition— Comprehensive Tutorial 3 1 Searching the Web Using Search Engines and Directories Effectively Tutorial.
QUESTION AND ANSWERING. Overview What is Question Answering? Why use it? How does it work? Problems Examples Future.
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
Elaine Ménard & Margaret Smithglass School of Information Studies McGill University [Canada] July 5 th, 2011 Babel revisited: A taxonomy for ordinary images.
Higher English Close Reading Types of Questions Understanding Questions Tuesday 8 OctoberCMCM1.
Evaluating Semantic Metadata without the Presence of a Gold Standard Yuangui Lei, Andriy Nikolov, Victoria Uren, Enrico Motta Knowledge Media Institute,
A Novel Pattern Learning Method for Open Domain Question Answering IJCNLP 2004 Yongping Du, Xuanjing Huang, Xin Li, Lide Wu.
BioRAT: Extracting Biological Information from Full-length Papers David P.A. Corney, Bernard F. Buxton, William B. Langdon and David T. Jones Bioinformatics.
Evaluation of (Search) Results How do we know if our results are any good? Evaluating a search engine  Benchmarks  Precision and recall Results summaries:
21/11/20151Gianluca Demartini Ranking Clusters for Web Search Gianluca Demartini Paul–Alexandru Chirita Ingo Brunkhorst Wolfgang Nejdl L3S Info Lunch Hannover,
Department of Software and Computing Systems Research Group of Language Processing and Information Systems The DLSIUAES Team’s Participation in the TAC.
Summarizing Encyclopedic Term Descriptions on the Web from Coling 2004 Atsushi Fujii and Tetsuya Ishikawa Graduate School of Library, Information and Media.
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
Teaching Writing.
A Classification-based Approach to Question Answering in Discussion Boards Liangjie Hong, Brian D. Davison Lehigh University (SIGIR ’ 09) Speaker: Cho,
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.
Acquisition of Categorized Named Entities for Web Search Marius Pasca Google Inc. from Conference on Information and Knowledge Management (CIKM) ’04.
Word Create a basic TOC. Course contents Overview: table of contents basics Lesson 1: About tables of contents Lesson 2: Format your table of contents.
The Cross Language Image Retrieval Track: ImageCLEF Breakout session discussion.
Exploiting Named Entity Taggers in a Second Language Thamar Solorio Computer Science Department National Institute of Astrophysics, Optics and Electronics.
Pastra and Saggion, EACL 2003 Colouring Summaries BLEU Katerina Pastra and Horacio Saggion Department of Computer Science, Natural Language Processing.
An evolutionary approach for improving the quality of automatic summaries Constantin Orasan Research Group in Computational Linguistics School of Humanities,
1 Question Answering and Logistics. 2 Class Logistics  Comments on proposals will be returned next week and may be available as early as Monday  Look.
How to Turnitin Dr Stephen Rankin Lecturer in Academic Writing and Literacy Murdoch University A 6 step guide for submitting your assignments to Turnitin.
Using Human Language Technology for Automatic Annotation and Indexing of Digital Library Content Kalina Bontcheva, Diana Maynard, Hamish Cunningham, Horacio.
Certification of Reusable Software Artifacts
Designing Cross-Language Information Retrieval System using various Techniques of Query Expansion and Indexing for Improved Performance  Hello everyone,
Social Knowledge Mining
Presentation 王睿.
Instructional Learning Cycle:
Workforce Engagement Survey
What is the Entrance Exams Task
Retrieval Evaluation - Reference Collections
Presentation transcript:

AnswerFinder Question Answering from your Desktop Mark A. Greenwood Natural Language Processing Group Department of Computer Science University of Sheffield, UK

7th of January 2004CLUK, 7th Annual Research Colloquium Outline of Talk What is Question Answering?  Different Question Types  A Generic Question Answering Framework  Evaluating Question Answering Systems System Description  Question Typing  Information Retrieval  Locating Possible Answers  A Detailed Example Results and Evaluation Desktop Question Answering  A Brief Comparison to Other On-Line Question Answering Systems Conclusions and Future Work

7th of January 2004CLUK, 7th Annual Research Colloquium What is Question Answering? The main aim of QA is to present the user with a short answer to a question rather than a list of possibly relevant documents. As it becomes more and more difficult to find answers on the WWW using standard search engines, question answering technology will become increasingly important. Answering questions using the web is already enough of a problem for it to appear in fiction (Marshall, 2002): “I like the Internet. Really, I do. Any time I need a piece of shareware or I want to find out the weather in Bogotá… I’m the first guy to get the modem humming. But as a source of information, it sucks. You got a billion pieces of data, struggling to be heard and seen and downloaded, and anything I want to know seems to get trampled underfoot in the crowd.”

7th of January 2004CLUK, 7th Annual Research Colloquium Different Question Types Clearly there are many different types of questions:  When was Mozart born? Question requires a single fact as an answer. Answer may be found verbatim in text i.e. “Mozart was born in 1756”.  How did Socrates die? Finding an answer may require reasoning. In this example die has to be linked with drinking poisoned wine.  How do I assemble a bike? The full answer may require fusing information from many different sources. The complexity can range from simple lists to script-based answers.  Is the Earth flat? Requires a simple yes/no answer. The systems outlined in this presentation attempt to answer the first two types of question.

7th of January 2004CLUK, 7th Annual Research Colloquium A Generic QA Framework A search engine is used to find the n most relevant documents in the document collection. These documents are then processed with respect to the question to produce a set of answers which are passed back to the user. Most of the differences between question answering systems are centred around the document processing stage.

7th of January 2004CLUK, 7th Annual Research Colloquium Evaluating QA Systems The biggest independent evaluations of question answering systems have been carried out at TREC (Text Retrieval Conference) over the past five years.  Five hundred factoid questions are provided and the groups taking part have a week in which to process the questions and return one answer per question.  No changes to systems are allowed between the time the questions are received and the time at which the answers are submitted. Not only do these annual evaluations give groups a chance to see how their systems perform against those from other institutions but more importantly it is slowly building an invaluable collection of resources, including questions and their associated answers, which can be used for further development and testing. Different metrics have been used over the years but the current metric is simply the percentage of questions correctly answered.

7th of January 2004CLUK, 7th Annual Research Colloquium Outline of Talk What is Question Answering?  Different Question Types  A Generic Question Answering Framework  Evaluating Question Answering Systems System Description  Question Typing  Information Retrieval  Locating Possible Answers  A Detailed Example Results and Evaluation Desktop Question Answering  A Brief Comparison to Other On-Line Question Answering Systems Conclusions and Future Work

7th of January 2004CLUK, 7th Annual Research Colloquium System Description Many of the systems which have proved successful in previous TREC evaluations have made use of a fine grained set of answer types.  One system (Harabagiu et al., 2000) has an answer type DOG BREED  The answer topology described in (Hovy et al., 2000) contains 94 different answer types. The original idea behind building the QA system underlying AnswerFinder was to determine how well a system which used only a fine grained set of answer types could perform.  The completed system consists of three distinct phases: Question Typing Information Retrieval Locating Possible Answers

7th of January 2004CLUK, 7th Annual Research Colloquium Question Typing The first stage of processing is to determine the semantic type of the expected answer. The semantic type, S, is determined through rules which examine the question, Q:  If Q contains ‘congressman’ and does not start with ‘where’ or ‘when’ then S is person:male  If Q contains ‘measured in’ then S is measurement_unit  If Q contains ‘univesity’ and does not start with ‘who’, ‘where’ or ‘when’ then S is organization  If Q contains ‘volcano’ and does not start with ‘who’ or ‘when’ then S is location The current system includes rules which can detect 46 different answer types.

7th of January 2004CLUK, 7th Annual Research Colloquium Information Retrieval This is by far the simplest part of the question answering system with the question being passed, as is, to an appropriate search engine:  Okapi is used, to search the A QUAINT collection, when answering the TREC questions.  XXXXXXXX is used, to search the Internet, when using AnswerFinder as a general purpose question answering system. The top n relevant documents, as determined by the search engine, are then retrieved ready for the final processing stage.

7th of January 2004CLUK, 7th Annual Research Colloquium Locating Possible Answers The only answers we attempt to locate are entities which the system can recognise. Locating possible answers consists therefore of extracting all entities of the required type from the relevant documents.  Entities are currently extracted using modified versions of the gazetteer lists and named entity transducer supplied with the GATE 2 framework (Cunningham et al., 2002). All entities of the correct type are retained as possible answers unless they fail one or both of the following tests:  The document the current entity appears in must contain all the entities in the question.  A possible answer entity must not contain any of the question words (ignoring stopwords).

7th of January 2004CLUK, 7th Annual Research Colloquium Locating Possible Answers All the remaining entities are then grouped together using the following equivalence test (Brill et al., 2001): Two answers are said to be equivalent if all of the non-stopwords in one are present in the other or vice versa. The resulting answer groups are then ordered by:  the frequency of occurrence of all answers within the group  the highest ranked document in which an answer in the group appears. This sorted list (or the top n answers) is then presented, along with a supporting snippet, to the user of the system.

7th of January 2004CLUK, 7th Annual Research Colloquium A Detailed Example Q: How high is Everest?D 1 : Everest’s 29,035 feet is 5.4 miles above sea level… D 2 : At 29,035 feet the summit of Everest is the highest… If Q contains ‘how’ and ‘high’ then the semantic class, S, is measurement:distance 29,035 feet measurement:distance(‘5.4 miles’) 1 measurement:distance(‘29,035 feet’) 2 location(‘Everest’) 2 Known Entities#

7th of January 2004CLUK, 7th Annual Research Colloquium Outline of Talk What is Question Answering?  Different Question Types  A Generic Question Answering Framework  Evaluating Question Answering Systems System Description  Question Typing  Information Retrieval  Locating Possible Answers  A Detailed Example Results and Evaluation Desktop Question Answering  A Brief Comparison to Other On-Line Question Answering Systems Conclusions and Future Work

7th of January 2004CLUK, 7th Annual Research Colloquium Results and Evaluation The underlying system was tested over the 500 factoid questions used in TREC 2002 (Voorhees, 2002): Results for the question typing stage were as follows:  16.8% (84/500) of the questions were of an unknown type and hence could never be answered correctly.  1.44% (6/416) of those questions which were typed were given the wrong type and hence could never be answered correctly.  Therefore the maximum attainable score of the entire system, irrespective of any future processing, is 82% (410/500). Results for the information retrieval stage were as follows:  At least one relevant document was found for 256 of the of the correctly typed questions.  Therefore the maximum attainable score of the entire system, irrespective of further processing, is 51.2% (256/500).

7th of January 2004CLUK, 7th Annual Research Colloquium Results and Evaluation Results for the question answering stage were as follows:  25.6% (128/500) questions were correctly answered by the system using this approach. These results are not overly impressive especially when compared with the best performing systems which can answer approximately 85% of the same five hundred questions (Moldovan et al, 2002). Users of web search engines are, however, used to looking at a set of relevant documents and so would probably be happy looking at a handful of short answers.  If we examine the top five answers returned for each question then the system correctly answers 35.8% (179/500) of the questions which is 69.9% (179/256) of the maximum attainable score.  If we examine all the answers returned for each question then 38.6% (193/500) of the questions are correctly answered which is 75.4% (193/256) of the maximum attainable score, but this involves displaying over 20 answers per question.

7th of January 2004CLUK, 7th Annual Research Colloquium Outline of Talk What is Question Answering?  Different Question Types  A Generic Question Answering Framework  Evaluating Question Answering Systems System Description  Question Typing  Information Retrieval  Locating Possible Answers  A Detailed Example Results and Evaluation Desktop Question Answering  A Brief Comparison to Other On-Line Question Answering Systems Conclusions and Future Work

7th of January 2004CLUK, 7th Annual Research Colloquium Desktop Question Answering Question answering may be an interesting research topic but what is needed is an application that is as simple to use as a modern web search engine.  No training or special knowledge required to use.  Must respond within a reasonable period of time.  Answers should be exact but should also be supported by a small snippet of text so that users don’t have to read the supporting document to verify the answer AnswerFinder attempts to meet all of these requirements…

7th of January 2004CLUK, 7th Annual Research Colloquium Desktop Question Answering When was Gustav Holst born?

7th of January 2004CLUK, 7th Annual Research Colloquium Brief Comparison - PowerAnswer PowerAnswer is developed by the team responsible for the best performing TREC system. At TREC 2002 their entry answered approx. 85% of the questions. Unfortunately PowerAnswer acts more like a search engine than a question answering system:  Each answer is a sentence or long phrase  No attempt is made to cluster/remove sentences which contain the same answers This is strange as TREC results show that this system is very good at finding a single exact answer to a question. … and get the answer!

7th of January 2004CLUK, 7th Annual Research Colloquium Brief Comparison - AnswerBus Very similar to PowerAnswer in that:  The answers presented are full sentences.  No attempt is made to cluster/remove sentences containing the same answer. The interesting thing to note about AnswerBus is that questions can be asked in more than one language; English, French, Spanish, German, Italian or Portuguese – although all answers are given in English. The developer claims the system answers 70.5% of the TREC 8 questions, although  The TREC 8 question set is not a good reflection of real world questions,  Finding exact answers, as the TREC evaluations have shown, is a harder task than simply finding answer bearing sentences.

7th of January 2004CLUK, 7th Annual Research Colloquium Brief Comparison - XXXX The NSIR system, from the University of Michigan, is much closer to AnswerFinder than PowerAnswer or AnswerBus:  Uses standard web search engines to find relevant documents.  Returns a list of ranked exact answers. Unfortunately no context or confidence level is given for each answer so users would still have to refer to the relevant documents to verify that a given answer is correct. NSIR was entered in TREC 2002 correctly answering 24.2% of the questions.  Very similar to the 25.6% obtained by AnswerFinder over the same question set.

7th of January 2004CLUK, 7th Annual Research Colloquium Brief Comparison - IONAUT IONAUT is the system most close to AnswerFinder when viewed from the user’s perspective.  A ranked list of answers is presented.  Supporting snippets of context are also displayed. Unfortunately the exact answers are not linked to specific snippets, so it is not immediately clear which snippet supports which answer.  This problem is compounded by the fact that multiple snippets may support a single answer as no attempt has been made to cluster/remove snippets which support the same answer.

7th of January 2004CLUK, 7th Annual Research Colloquium Outline of Talk What is Question Answering?  Different Question Types  A Generic Question Answering Framework  Evaluating Question Answering Systems System Description  Question Typing  Information Retrieval  Locating Possible Answers  A Detailed Example Results and Evaluation Desktop Question Answering  A Brief Comparison to Other On-Line Question Answering Systems Conclusions and Future Work

7th of January 2004CLUK, 7th Annual Research Colloquium Conclusions The original aim in developing the underlying question answering system was to determine how well only a fine grained system of answer types would perform.  The system answers approximately 26% of the TREC 11 questions.  The average performance by participants in TREC 11 was 22%.  The best performing system at TREC 11 scored approximately 85%. The aim of developing AnswerFinder was to provide access to question answering technology in a manner similar to current web search engines.  An interface similar to a web browser is used to both enter the question and to display the answers.  The answers are displayed in a similar fashion to standard web search results.  Very little extra time is required to locate possible answers over and above simply collecting the relevant documents.

7th of January 2004CLUK, 7th Annual Research Colloquium Future Work The question typing stage could be improved through either the edition of more rules or by replacing the rules with an automatically acquired classifier (Li and Roth, 2002). It should be clear that increasing the types of entities we can recognise will increase the percentage of questions we can answer. Unfortunately this is a task that is both time- consuming and never-ending. A possible extension to this approach is to include answer extraction patterns (Greenwood and Gaizauskas, 2003).  These patterns are enhanced regular expressions in which certain tags will match multi-word terms.  For example questions such as “What does CPR stand for?” generate patterns such as “ NounChunK ( X ) ” where CPR is substituted for X to select a noun chunk that will be suggested as a possible answer.

Any Questions? Copies of these slides can be found at: AnswerFinder can be downloaded from:

7th of January 2004CLUK, 7th Annual Research Colloquium Bibliography Eric Brill, Jimmy Lin, Michele Banko, Susan Dumais and Andrew Ng. Data-Intensive Question Answering. In Proceedings of the 10th Text REtrieval Conference, Hamish Cunningham, Diana Maynard, Kalina Bontcheva and Valentin Tablan. GATE: A framework and graphical development environment for robust NLP tools and applications. In Proceedings of the 40th Anniversary Meeting of the Association for Computational Linguistics, Mark A. Greenwood and Robert Gaizauskas. Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering. In Proceedings of the Workshop on Natural Language Processing for Question Answering (EACL03), pages 29–34, Budapest, Hungary, April 14, Sanda Harabagiu, Dan Moldovan, Marius Paşca, Rada Mihalcea, Mihai Surdeanu, Răzvan Bunescu, Roxana Gîrju, Vasile Rus and Paul Morărescu. FALCON: Boosting Knowledge for Answer Engines. In Proceedings of the 9th Text REtrieval Conference, Eduard Hovy, Laurie Gerber, Ulf Hermjakob, Michael Junk and Chin-Yew Lin. Question Answering in Webclopedia. In Proceedings of the 9th Text REtrieval Conference, Xin Li and Dan Roth. Learning Question Classifiers. In Proceedings of the 19 th International Conference on Computational Linguistics (COLING’02), Michael Marshall. The Straw Men. HarperCollins Publishers, Dan Moldovan, Sanda Harabagiu, Roxana Girju, Paul Morarescu, Finley Lacatusu, Adrian Novischi, Adriana Badulescu, and Orest Bolohan. LCC Tools for Question Answering. In Proceedings of the 11th Text REtrieval Conference, Ellen M. Voorhees. Overview of the TREC 2002 Question Answering Track. In Proceedings of the 11th Text REtrieval Conference, 2002.