SEMANTIC VERIFICATION IN AN ONLINE FACT SEEKING ENVIRONMENT DMITRI ROUSSINOV, OZGUR TURETKEN Speaker: Li, HueiJyun Advisor: Koh, JiaLing Date: 2008/5/1.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Date: 2013/1/17 Author: Yang Liu, Ruihua Song, Yu Chen, Jian-Yun Nie and Ji-Rong Wen Source: SIGIR12 Advisor: Jia-ling Koh Speaker: Chen-Yu Huang Adaptive.
Learning to Suggest: A Machine Learning Framework for Ranking Query Suggestions Date: 2013/02/18 Author: Umut Ozertem, Olivier Chapelle, Pinar Donmez,
Improvements and extras Paul Thomas CSIRO. Overview of the lectures 1.Introduction to information retrieval (IR) 2.Ranked retrieval 3.Probabilistic retrieval.
Document Summarization using Conditional Random Fields Dou Shen, Jian-Tao Sun, Hua Li, Qiang Yang, Zheng Chen IJCAI 2007 Hao-Chin Chang Department of Computer.
Mustafa Cayci INFS 795 An Evaluation on Feature Selection for Text Clustering.
SEARCHING QUESTION AND ANSWER ARCHIVES Dr. Jiwoon Jeon Presented by CHARANYA VENKATESH KUMAR.
Sequence Clustering and Labeling for Unsupervised Query Intent Discovery Speaker: Po-Hsien Shih Advisor: Jia-Ling Koh Source: WSDM’12 Date: 1 November,
Named Entity Recognition in Query Jiafeng Guo, Gu Xu, Xueqi Cheng, Hang Li (ACM SIGIR 2009) Speaker: Yi-Lin,Hsu Advisor: Dr. Koh, Jia-ling Date: 2009/11/16.
DOMAIN DEPENDENT QUERY REFORMULATION FOR WEB SEARCH Date : 2013/06/17 Author : Van Dang, Giridhar Kumaran, Adam Troy Source : CIKM’12 Advisor : Dr. Jia-Ling.
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
1 Entity Ranking Using Wikipedia as a Pivot (CIKM 10’) Rianne Kaptein, Pavel Serdyukov, Arjen de Vries, Jaap Kamps 2010/12/14 Yu-wen,Hsu.
ITCS 6010 Natural Language Understanding. Natural Language Processing What is it? Studies the problems inherent in the processing and manipulation of.
Presented by Zeehasham Rasheed
J. Chen, O. R. Zaiane and R. Goebel An Unsupervised Approach to Cluster Web Search Results based on Word Sense Communities.
Overview of Search Engines
Finding Advertising Keywords on Web Pages Scott Wen-tau YihJoshua Goodman Microsoft Research Vitor R. Carvalho Carnegie Mellon University.
Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification on Reviews Peter D. Turney Institute for Information Technology National.
1 Opinion Spam and Analysis (WSDM,08)Nitin Jindal and Bing Liu Date: 04/06/09 Speaker: Hsu, Yu-Wen Advisor: Dr. Koh, Jia-Ling.
Web Usage Mining with Semantic Analysis Date: 2013/12/18 Author: Laura Hollink, Peter Mika, Roi Blanco Source: WWW’13 Advisor: Jia-Ling Koh Speaker: Pei-Hao.
1 Context-Aware Search Personalization with Concept Preference CIKM’11 Advisor : Jia Ling, Koh Speaker : SHENG HONG, CHUNG.
Reyyan Yeniterzi Weakly-Supervised Discovery of Named Entities Using Web Search Queries Marius Pasca Google CIKM 2007.
AnswerBus Question Answering System Zhiping Zheng School of Information, University of Michigan HLT 2002.
Michael Cafarella Alon HalevyNodira Khoussainova University of Washington Google, incUniversity of Washington Data Integration for Relational Web.
Ontology-Driven Automatic Entity Disambiguation in Unstructured Text Jed Hassell.
CIKM’09 Date:2010/8/24 Advisor: Dr. Koh, Jia-Ling Speaker: Lin, Yi-Jhen 1.
A Probabilistic Graphical Model for Joint Answer Ranking in Question Answering Jeongwoo Ko, Luo Si, Eric Nyberg (SIGIR ’ 07) Speaker: Cho, Chin Wei Advisor:
Understanding and Predicting Personal Navigation Date : 2012/4/16 Source : WSDM 11 Speaker : Chiu, I- Chih Advisor : Dr. Koh Jia-ling 1.
Natural Language Based Reformulation Resource and Web Exploitation for Question Answering Ulf Hermjakob, Abdessamad Echihabi, Daniel Marcu University of.
Péter Schönhofen – Ad Hoc Hungarian → English – CLEF Workshop 20 Sep 2007 Performing Cross-Language Retrieval with Wikipedia Participation report for Ad.
Retrieval Models for Question and Answer Archives Xiaobing Xue, Jiwoon Jeon, W. Bruce Croft Computer Science Department University of Massachusetts, Google,
A Language Independent Method for Question Classification COLING 2004.
21/11/2002 The Integration of Lexical Knowledge and External Resources for QA Hui YANG, Tat-Seng Chua Pris, School of Computing.
Presenter: Shanshan Lu 03/04/2010
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
A Novel Pattern Learning Method for Open Domain Question Answering IJCNLP 2004 Yongping Du, Xuanjing Huang, Xin Li, Lide Wu.
Understanding User’s Query Intent with Wikipedia G 여 승 후.
A Scalable Machine Learning Approach for Semi-Structured Named Entity Recognition Utku Irmak(Yahoo! Labs) Reiner Kraft(Yahoo! Inc.) WWW 2010(Information.
Qi Guo Emory University Ryen White, Susan Dumais, Jue Wang, Blake Anderson Microsoft Presented by Tetsuya Sakai, Microsoft Research.
Algorithmic Detection of Semantic Similarity WWW 2005.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
LOGO 1 Corroborate and Learn Facts from the Web Advisor : Dr. Koh Jia-Ling Speaker : Tu Yi-Lang Date : Shubin Zhao, Jonathan Betz (KDD '07 )
LING 573 Deliverable 3 Jonggun Park Haotian He Maria Antoniak Ron Lockwood.
Probabilistic Latent Query Analysis for Combining Multiple Retrieval Sources Rong Yan Alexander G. Hauptmann School of Computer Science Carnegie Mellon.
Summarizing Encyclopedic Term Descriptions on the Web from Coling 2004 Atsushi Fujii and Tetsuya Ishikawa Graduate School of Library, Information and Media.
A Word Clustering Approach for Language Model-based Sentence Retrieval in Question Answering Systems Saeedeh Momtazi, Dietrich Klakow University of Saarland,Germany.
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
Automatic Question Answering  Introduction  Factoid Based Question Answering.
A Classification-based Approach to Question Answering in Discussion Boards Liangjie Hong, Brian D. Davison Lehigh University (SIGIR ’ 09) Speaker: Cho,
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
Date: 2013/6/10 Author: Shiwen Cheng, Arash Termehchy, Vagelis Hristidis Source: CIKM’12 Advisor: Jia-ling Koh Speaker: Chen-Yu Huang Predicting the Effectiveness.
Improved Video Categorization from Text Metadata and User Comments ACM SIGIR 2011:Research and development in Information Retrieval - Katja Filippova -
Date: 2012/11/29 Author: Chen Wang, Keping Bi, Yunhua Hu, Hang Li, Guihong Cao Source: WSDM’12 Advisor: Jia-ling, Koh Speaker: Shun-Chen, Cheng.
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
Ranking of Database Query Results Nitesh Maan, Arujn Saraswat, Nishant Kapoor.
CONTEXTUAL SEARCH AND NAME DISAMBIGUATION IN USING GRAPHS EINAT MINKOV, WILLIAM W. COHEN, ANDREW Y. NG SIGIR’06 Date: 2008/7/17 Advisor: Dr. Koh,
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Date: 2013/9/25 Author: Mikhail Ageev, Dmitry Lagun, Eugene Agichtein Source: SIGIR’13 Advisor: Jia-ling Koh Speaker: Chen-Yu Huang Improving Search Result.
Predicting Short-Term Interests Using Activity-Based Search Context CIKM’10 Advisor: Jia Ling, Koh Speaker: Yu Cheng, Hsieh.
Question Answering Passage Retrieval Using Dependency Relations (SIGIR 2005) (National University of Singapore) Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan,
Multi-Class Sentiment Analysis with Clustering and Score Representation Yan Zhu.
Short Text Similarity with Word Embedding Date: 2016/03/28 Author: Tom Kenter, Maarten de Rijke Source: CIKM’15 Advisor: Jia-Ling Koh Speaker: Chih-Hsuan.
Search Engine Architecture
Erasmus University Rotterdam
Martin Rajman, Martin Vesely
Multimedia Information Retrieval
Social Knowledge Mining
Presentation transcript:

SEMANTIC VERIFICATION IN AN ONLINE FACT SEEKING ENVIRONMENT DMITRI ROUSSINOV, OZGUR TURETKEN Speaker: Li, HueiJyun Advisor: Koh, JiaLing Date: 2008/5/1 1

OUTLINE Introduction System Overview Automated Semantic Verification Approach Empirical Evaluation Conclusions, Limitations and Future Research 2

INTRODUCTION The goal of Question Answering (QA) is to locate, extract, and represent a specific answer to a user question expressed in a natural language Answers to many natural language questions are expected to belong to a certain semantic category E.g. What color is the sky? Difficult for the current QA systems since the correct answer is not guaranteed to be found in an explicit form such as in the sentence The color of the sky is blue May need to be extracted from a sentence answering it implicitly, e.g. such as I saw a vast blue sky above me 3

INTRODUCTION The current popular approach to solving this “semantic” matching problem is through developing an extensive taxonomy of possible semantic categories Requires the anticipation of all possible questions and substantial manual effort Works relatively well with more common categories E.g., cities, countries, organizations, writers, musicians, etc. Handling more rare categories is still an challenge What was the name of the first Russian astronaut to do a spacewalk? What soft drink contains the largest amount of caffeine? 4

SYSTEM OVERVIEW The general idea: to apply pattern matching and take the advantage of the redundancy on the web What is the capital of Taiwan? The capital of Taiwan is Taipei. \A matches a pattern \Q \Q: the question part (“ The capital of Taiwan ”) \A: the text that forms a candidate answer (“ Taipei ”) 5

SYSTEM OVERVIEW Automatically create and train up to 50 patterns for each question type, based on a training data set consisting of open domain QA pairs Question type: what is, what was, where is, how far, how many, how big etc. Open domain QA pairs: those available from past TREC conference Through training, each pattern is assigned a probability that the matching text contains the correct answer This probability is used in the triangulation (confirming/disconfirming) process that re-ranks candidate answers 6

SYSTEM OVERVIEW In which city is Eiffel Tower located? Type identification: In which \T is \Q \V Query Modulation (QM): converts each answer pattern into a query for a General Purpose Search Engine (GPSE) E.g., \Q is \V in \A The Google query would be + “ Eiffel Tower is located in ” 7 “ city ” The semantic category of the expected answer “ Eiffel Tower ” Question part “ located ” Verb

SYSTEM OVERVIEW Answer Matching: “ Eiffel Tower is located in the centre of Paris, the capital of France ” Candidate answer: “ the centre of Paris, the capital of France ”, with a corresponding probability of containing a correct answer (i.e. 0.5) obtained previously for each pattern by training Answer Detailing (AD): produces more candidate answers by forming sub-phrase from original candidate answers that: 1. Do not exceed 3 words (not counting “stop words” such as a, the, in, on ) 2. Do not cross punctuation marks 3. Do not start or end with stop words Centre, Paris, capital, France, centre of Paris, capital of France 8

SYSTEM OVERVIEW Semantic Adjustment: use a small set of independently considered semantic boolean (yes/no) features of candidate answers is number, is date, is year, is location, is person name, is proper name Not perfectly accurate, those features are detected by matching simple hand-crafted regular expressions 9

SYSTEM OVERVIEW Triangulation: the candidate answers are triangulated (confirmed or disconfirmed) against each other and then reordered according to their final score Promotes those candidate answers that are repeated the most Eventually assigns a score s c in the [0, 1] range which can be informally interpreted as the estimate of the probability of the candidate answer being correct These probabilities are independent in the sense that they do not necessarily add up to 1 across all the candidate answer since multiple correct answers are indeed possible 10

SYSTEM OVERVIEW Semantic Verification: If the answer to a question is expected to belong to a certain semantic category (e.g., City ), then candidate answers are re-ordered according to their semantic verification scores 11

AUTOMATED SEMANTIC VERIFICATION APPROACH Intuition behind the approach studied: What soft drink contains the largest amount of caffeine? Expect a candidate answer (e.g. pepsi ) to belong to a specific category ( soft drink ) Implies that one of the possible answers to the question What is pepsi? should be soft drink Ask this question automatically, perform the same necessary steps, and check the certain number of top answers to see if they contain the desired category 12

AUTOMATED SEMANTIC VERIFICATION APPROACH Select 16 most accurate patterns from those automatically identified and trained for “what is” type question For each candidate answer, system queries the underlying web search engine only for the total number of pages that match the modulated query for each pattern For example, “pepsi is a soft drink”, matches 127 pages from Google; “pepsi is a type of a soft drink” matches 0, etc Denote the corresponding number of match as m 1 …m 16, and the aggregate total number of matches is denoted as M 13

AUTOMATED SEMANTIC VERIFICATION APPROACH The role of semantic verification in the QA process: The purpose of semantic verification is to estimate the probability of the given candidate answer to belong to the specified category For example, since pepsi is a soft drink, that probability should be close to 1 In some less clear cases (e.g. mosquito is an animal ) it can be further from 1 14

AUTOMATED SEMANTIC VERIFICATION APPROACH Since this approach draws on the aggregated opinions of all of the Web contributors, it may produce category memberships that are considered wrong according to common myths, misconceptions, or even jokes, which are frequently circulating on the Web For example: “coffee is a soft drink” It is important to aggregate across multiple pattern matches It is also important to consider the overall statistics of both the category and the candidate answer For example, the total numbers of web pages that contain them Obviously, the more frequent candidate answers need to be demoted 15

AUTOMATED SEMANTIC VERIFICATION APPROACH Combine all these variables into one model that estimates the probability of membership in a given semantic category  s The final score according to which the candidate answers are sorted before being presented to the user can be computed as: s f = s ‧ s c s c : the “initial score” (probability of being correct) of the candidate answer after all the processing steps before semantic verification is applied 16

AUTOMATED SEMANTIC VERIFICATION APPROACH Building the predictive model: using binary logistic regression techniques to compute s Models simply used different combinations of the following variables may help to discriminate between the semantic match and mis-match m 1, m 2, …, m 16 : match for each of the patterns M = m 1 + m 2 + …+ m 16 : total number of pattern matches DF A : number of pages on the web containing the candidate answer (e.g. pepsi ) DF C : number of pages on the web containing the expected category (e.g. soft drink ) DF AC : number of web pages containing both the candidate answer (e.g. pepsi ) and the expected category (e.g. soft drink ) 17

AUTOMATED SEMANTIC VERIFICATION APPROACH Models: Model 0) s = const, or simply no semantic verification Model 1) s = s(m 1, m 2, …, m 16, DF A, DF C, DF AC ) : all original variables in their non-transformed form Model 2) s = s(log(M+1), log(DF A ), log(DF C ), log(DF AC )) : all variables are log-transformed, aggregate variable M is used instead of separate m-s Model 3) s = s(m 1, m 2, …, m 16, log(DF A ), log(DF C ), log(DF AC )): only the document frequencies are log- transformed and used independently Model 4) s = s(log(DF A ), log(DF C ), log(DF AC ), log(M+1)) : only the total number of pattern matches is log-transformed, no separate m-s is used Model 5) s = s(log(DF A ), log(DF C ), log(M+1)) : same as 4 but without log(DF AC ) Model 6) s = s(log(DF AC, log(M+1)) : only DF AC (co- occurrence) and log of M – the simplest model 18

EMPIRICAL EVALUATION Data sets In order to train the model, creates a labeled data set: Took all the 68 questions from TREC 2003 that specified the semantic category explicitly Run the questions through QA system without semantic verification and manually labeled the correctness of the semantic category as true/false for top 30 candidate answers for each question of 68 questions 19

EMPIRICAL EVALUATION Category Prediction 20

EMPIRICAL EVALUATION Improving Answering Accuracy 21 MRR: Mean reciprocal rank TRDR: Mean total reciprocal rank

EMPIRICAL EVALUATION Testing on a Previously Unseen Set 22

CONCLUSIONS, LIMITATION AND FUTURE RESEARCH Have introduced a novel semantic verification algorithm and demonstrated that it improves the accuracy of answers produced by a redundancy- based question answering system Limitation 23