Presentation 151220111 王睿.

Slides:



Advertisements
Similar presentations
Leveraging Community-built Knowledge For Type Coercion In Question Answering Aditya Kalyanpur, J William Murdock, James Fan and Chris Welty Mehdi AllahyariSpring.
Advertisements

Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
Writing Formal Reports
CS 330 Programming Languages 09 / 13 / 2007 Instructor: Michael Eckmann.
Computer comunication B Information retrieval. Information retrieval: introduction 1 This topic addresses the question on how it is possible to find relevant.
UNDERSTANDING BILINGUAL TRANSLATION OF SPECIALIZED TEXTS.
Enhance legal retrieval applications with an automatically induced knowledge base Ka Kan Lo.
Essay Writing What makes a good essay?. Essay Writing What is a good essay? Planning Essay structure Editing and proofreading Referencing and avoiding.
An Automatic Segmentation Method Combined with Length Descending and String Frequency Statistics for Chinese Shaohua Jiang, Yanzhong Dang Institute of.
Longbiao Kang, Baotian Hu, Xiangping Wu, Qingcai Chen, and Yan He Intelligent Computing Research Center, School of Computer Science and Technology, Harbin.
OMAP: An Implemented Framework for Automatically Aligning OWL Ontologies SWAP, December, 2005 Raphaël Troncy, Umberto Straccia ISTI-CNR
Modeling (Chap. 2) Modern Information Retrieval Spring 2000.
Dr. Alireza Isfandyari-Moghaddam Department of Library and Information Studies, Islamic Azad University, Hamedan Branch
Computational Methods to Vocalize Arabic Texts H. Safadi*, O. Al Dakkak** & N. Ghneim**
C OLLECTIVE ANNOTATION OF WIKIPEDIA ENTITIES IN WEB TEXT - Presented by Avinash S Bharadwaj ( )
IMSS005 Computer Science Seminar
Exploiting Wikipedia as External Knowledge for Document Clustering Sakyasingha Dasgupta, Pradeep Ghosh Data Mining and Exploration-Presentation School.
Purpose of study A high-quality computing education equips pupils to use computational thinking and creativity to understand and change the world. Computing.
Attribute Extraction and Scoring: A Probabilistic Approach Taesung Lee, Zhongyuan Wang, Haixun Wang, Seung-won Hwang Microsoft Research Asia Speaker: Bo.
What should I know about the new TCAP/EOC? Testing Window
CONSTRUCTING OBJECTIVE TEST ITEMS: MULTIPLE-CHOICE FORMS CONSTRUCTING OBJECTIVE TEST ITEMS: MULTIPLE-CHOICE FORMS CHAPTER 8 AMY L. BLACKWELL JUNE 19, 2007.
1 Formal Models for Expert Finding on DBLP Bibliography Data Presented by: Hongbo Deng Co-worked with: Irwin King and Michael R. Lyu Department of Computer.
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
Péter Schönhofen – Ad Hoc Hungarian → English – CLEF Workshop 20 Sep 2007 Performing Cross-Language Retrieval with Wikipedia Participation report for Ad.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
BEHAVIORAL TARGETING IN ON-LINE ADVERTISING: AN EMPIRICAL STUDY AUTHORS: JOANNA JAWORSKA MARCIN SYDOW IN DEFENSE: XILING SUN & ARINDAM PAUL.
Wikipedia as Sense Inventory to Improve Diversity in Web Search Results Celina SantamariaJulio GonzaloJavier Artiles nlp.uned.es UNED,c/Juan del Rosal,
Understanding User’s Query Intent with Wikipedia G 여 승 후.
LOGO 1 Corroborate and Learn Facts from the Web Advisor : Dr. Koh Jia-Ling Speaker : Tu Yi-Lang Date : Shubin Zhao, Jonathan Betz (KDD '07 )
Department of Software and Computing Systems Research Group of Language Processing and Information Systems The DLSIUAES Team’s Participation in the TAC.
1 AC-Close: Efficiently Mining Approximate Closed Itemsets by Core Pattern Recovery Advisor : Dr. Koh Jia-Ling Speaker : Tu Yi-Lang Date : Hong.
1 Masters Thesis Presentation By Debotosh Dey AUTOMATIC CONSTRUCTION OF HASHTAGS HIERARCHIES UNIVERSITAT ROVIRA I VIRGILI Tarragona, June 2015 Supervised.
1Computer Sciences Department. Book: INTRODUCTION TO THE THEORY OF COMPUTATION, SECOND EDITION, by: MICHAEL SIPSER Reference 3Computer Sciences Department.
Format of Formal Reports
Toward Entity Retrieval over Structured and Text Data Mayssam Sayyadian, Azadeh Shakery, AnHai Doan, ChengXiang Zhai Department of Computer Science University.
CHAPTER 2 LITERATION REVIEW 1-1. LEARNING OUTCOMES 1.The reasons for a literature review being an essential part of every project. 2.The purpose of a.
 Criterion D: Knowledge and Understanding of Topic Studied  Criterion E: Reasoned Argument  Criterion F: Application of Analytical and Evaluative skills.
Dr. Marciano B. Melchor University of Ha’il, KINGDOM OF SAUDI ARABIA May 2013.
WP4 Models and Contents Quality Assessment
What is a CAT? What is a CAT?.
Designing Cross-Language Information Retrieval System using various Techniques of Query Expansion and Indexing for Improved Performance  Hello everyone,
Student Assessment Data Portal
Exploiting Wikipedia as External Knowledge for Document Clustering
G.W. CARVER MIDDLE EOC MATHEMATICS WORKSHOP December 7, 2016
Methods and Apparatus for Ranking Web Page Search Results
Quick Write Do you prefer fiction text vs. non-fiction text? Explain your answer.
Parts of an Academic Paper
CASE STUDY BY: JESSICA PATRON.
By Dr. Abdulrahman H. Altalhi
Data Structures Mohammed Thajeel To the second year students
Unit 4 Introducing the Study.
G.W. CARVER MIDDLE EOC MATHEMATICS WORKSHOP October 25, 2017
NURSING RESEARCH “THE LITERATURE REVIEW”
Location Recommendation — for Out-of-Town Users in Location-Based Social Network Yina Meng.
Key to Literature McKrell, Smith, Patrick.
LAW112 Assessment 3 Haley McEwen.
Software Requirements Specification Document
TECHNICAL REPORT.
A Framework for Benchmarking Entity-Annotation Systems
Block Matching for Ontologies
Writing reports Wrea Mohammed
Introduction to Information Retrieval
External assessment details—SL
Understanding Types of Formal Reports
Towards Solving Problems Using Textual Entailment over Identified Sources 2015/9/14 Weixi Zhu.
Discovering Companies we Know
AP U.S. History Exam Details
Precision and Recall.
Bug Localization with Combination of Deep Learning and Information Retrieval A. N. Lam et al. International Conference on Program Comprehension 2017.
Presentation transcript:

Presentation 151220111 王睿

Taking up the Gaokao Challenge : An Information Retrieval Approach Gong Cheng, Weixi Zhu, Ziwei Wang, Jianghui Cheng, Yuzhong Qu

Introduction Approach Retrieving pages Experiment Ranking and filtering pages Assessing options Experiment

Introduction A recently very hot AI challenge is to have the computer pass entrance examinations at different levels of education. Answering questions in a university’s entrance examination like Gaokao in China challenges AI technology. As a preliminary attempt to take up the Gaokao challenge, we aim to develop an approach to automatically answering multiple-choice questions in Gaokao. A related work of Project 863

Introduction Taking Wikipedia as the source of knowledge Proposing a three-stage approach Testing the approach on a set of real-life questions collected from recent history tests. Using strategies to assess options

Introduction To develop an approach to automatically answering multiple-choice questions in Gaokao. Complex stem background domain-specific expression

Approach A three-stage framework Recollecting relevant knowledge Filtering out non-essential knowledge Checking each option by some reasoning In our implementation of the framework, as shown in Fig. 2, (Chinese) Wikipedia is taken as the memory since it covers a wide range of knowledge and is freely accessible. Relevant knowledge is recollected by retrieving a set of pages from Wikipedia that match the question in two specific ways and thus may provide useful knowledge for answering the question. The subset of pages that are truly useful are regarded as evidence, which is drawn by ranking the retrieved pages according to multiple strategies for evaluating their usefulness, and filtering out bottom-ranked ones. The contents of the remaining pages will entail each option to some extent, based on which its truth is assessed.

Formal Description A multiple-choice question 𝑞 consists of a string 𝑠 𝑞 stem and a set of strings 𝑂 𝑞 called options; only one of 𝑂 𝑞 is the correct answer to 𝑠 𝑞 to be found.

Retrieving Pages Retrieving Concept Pages The description of a concept can be acquired from a Wikipedia page titled this concept, called a concept page. Matching the titles of Wikipedia pages to 𝑞 leftmost longest principle Some concepts mentioned in q can give clues about q; their descriptions may provide useful knowledge for answering q. The description of a concept can be acquired from a Wikipedia page titled this concept, called a concept page.

Retrieving Pages A retrieved concept page may belong to two categories of special pages in Wikipedia: disambiguation page and redirect page. Such a page needs to be replaced by another page as follows, to acquire the description of the desired concept. A concept name can be ambiguous. A retrieved concept page titled such an ambiguous concept name is usually a disambiguation page. It contains not the detailed description of any concept but links to other pages that describe concepts having the same name. Each link is accompanied by a short description of the corresponding concept for disambiguation purposes. A retrieved concept page can be a redirect page. Such a page simply forwards the reader to another page without providing any other information. Therefore, to acquire the description of the desired concept, a retrieved redirect page will be replaced by the page it is redirected to.

Retrieving Pages A retrieved disambiguation page: 𝑑𝑝 𝑐𝑛 An ambiguous concept name: 𝑐𝑛 A set of disambiguated concept pages: 𝑃 𝑐𝑛 The set of other page titles found in 𝑞 by Algorithm 1: 𝑃𝑇 𝑐𝑛 The abstract of a candidate concept page 𝑝∈𝑃 𝑐𝑛 : 𝑎𝑏𝑠𝑡𝑟 𝑝 The body of a candidate concept page 𝑝∈𝑃 𝑐𝑛 :𝑏𝑜𝑑𝑦 𝑝 The short description accompanying the link to 𝑝 in 𝑑𝑝 𝑐𝑛 : 𝑠𝑑 𝑝 The number of times a page title 𝑝𝑡 appears in text 𝑡: 𝑡𝑓 𝑝𝑡, 𝑡

Retrieving Pages The page 𝑝∈𝑃 𝑐𝑛 to replace 𝑝∈𝑃 𝑐𝑛 would have the highest ranking score: 𝑝𝑡∈𝑃𝑇 𝑐𝑛 𝛼⋅𝑡𝑓 𝑝𝑡,𝑎𝑏𝑠𝑡𝑟 𝑝 +𝛽⋅𝑡𝑓 𝑝𝑡,𝑏𝑜𝑑𝑦 𝑝 +𝛾⋅𝑡𝑓 𝑝𝑡,𝑠𝑑 𝑝 𝑝𝑡∈𝑃𝑇 𝑐𝑛 𝛼⋅𝑡𝑓 𝑝𝑡,𝑎𝑏𝑠𝑡𝑟 𝑝 +𝛽⋅𝑡𝑓 𝑝𝑡,𝑏𝑜𝑑𝑦 𝑝 +𝛾⋅𝑡𝑓 𝑝𝑡,𝑠𝑑 𝑝 where 𝛼,𝛽,𝛾∈ 0,1 are weights to be tuned, and are empirically set to 𝛼=0.8,𝛽=0.5,𝛾=1.0 in our experiment.

Retrieving Pages Retrieving Quote Pages Quotes in 𝑞 can give clues about 𝑞. They widely exist in Chinese, politics, geography, and history tests in Gaokao. However, quotes are often written in Classical Chinese, which is difficult to understand by speakers of modern Chinese and the computer because it uses different lexical items, and appears extremely concise and ambiguous. Apart from a quote itself, the context out of which the quote has been used may also provide useful knowledge for answering 𝑞. Wikipedia pages whose contents contain an exact match of the quote are called a quote page. A quote page that matches the largest (non-zero) number of these quotes is retrieved, denoted by 𝑃 𝑄 𝑆 𝑞 . Each option 𝑜∈ 𝑂 𝑞 is processed analogously, resulting in 𝑃 𝑄 𝑜 .

Ranking and Filtering Pages Let 𝑃 𝑅 𝑆 𝑞 , 𝑃 𝑅 𝑜 , 𝑃 𝑅 𝑂 𝑞 be the sets of concept and quote pages retrieved based on 𝑆 𝑞 , 𝑜 ∈𝑂 𝑞 and 𝑂 𝑞 , respectively: 𝑃 𝑅 𝑆 𝑞 = 𝑃 𝐶 𝑆 𝑞 ∪ 𝑃 𝑄 𝑆 𝑞 , 𝑃 𝑅 𝑜 = 𝑃 𝐶 𝑜 ∪ 𝑃 𝑄 𝑜 , 𝑃 𝑅 𝑂 𝑞 = ∪ 𝑜∈ 𝑂 𝑞 𝑃 𝑅 𝑜 Some of these pages are not useful for answering, but contain noise information. We develop three ranking strategies to filter them out: Centrality-based, domain-based, and relevance-based

Ranking and Filtering Pages Centrality-based Strategy Formally, each retrieved page 𝑝 𝑖 is represented by a vector 𝐹 𝑝 𝑖 , and is ranked by Domain-based Strategy When 𝑞 is known to be in a specific domain like history, Wikipedia’s category system can be used to filter out the retrieved pages not belonging to any historical categories. Historical categories consist of all the categories whose names contains the word ℎ𝑖𝑠𝑡𝑜𝑟𝑦, and their descendant categories. Relevance-based Strategy A retrieved page that is useful for answering 𝑞 should be relevant to both the stem and the option. Depending on the part of 𝑞 based on which 𝑝 𝑖 is retrieved, 𝑝 𝑖 is ranked by

Assessing Options Let 𝑃 𝐹 𝑆 𝑞 ⊆ 𝑃 𝑅 𝑆 𝑞 and 𝑃 𝐹 𝑜 ⊆ 𝑃 𝑅 𝑜 be the subsets of pages that remain after filtering by using a combination of the aforementioned strategies. We decompose the entailment into two parts: The extent to which a combination of 𝑆 𝑞 and 𝑜 can be entailed from 𝑆 𝑞 and 𝑃 𝐹 𝑆 𝑞 , and The extent to which a combination of 𝑆 𝑞 and 𝑜 can be entailed from 𝑜 and 𝑃 𝐹 𝑜 .

Experiments Dataset As a result, Metrics We evaluated our approach based on real-life questions appearing in recent history tests in Beijing. Specifically, 577 multiple-choice questions were collected. Three human experts were invited to collectively answer those questions only based on the contents of Chinese Wikipedia pages. As a result, 123 questions (21.32%), denoted by QS-A, could be successfully answered; 454 questions (78.68%), denoted by QS-B, went beyond the scope of Wikipedia, can could not be correctly answered without referring to the contents of history textbooks or other resources. Metrics Let G be the set of gold-standard pages used by the experts Let A be the set of pages found by an automatic approach. We measure the precision (P), recall (R), and F-score (F) of A:

Thanks Q & A