Hierarchical Summaries By: Dawn J. Lawrie University of Massachusetts, Amherst for Search.

Slides:



Advertisements
Similar presentations
Metadata in Carrot II Current metadata –TF.IDF for both documents and collections –Full-text index –Metadata are transferred between different nodes Potential.
Advertisements

Mustafa Cayci INFS 795 An Evaluation on Feature Selection for Text Clustering.
Language Models Naama Kraus (Modified by Amit Gross) Slides are based on Introduction to Information Retrieval Book by Manning, Raghavan and Schütze.
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
SEARCHING QUESTION AND ANSWER ARCHIVES Dr. Jiwoon Jeon Presented by CHARANYA VENKATESH KUMAR.
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Searchable Web sites Recommendation Date : 2012/2/20 Source : WSDM’11 Speaker : I- Chih Chiu Advisor : Dr. Koh Jia-ling 1.
Evaluating Search Engine
Query Operations: Automatic Local Analysis. Introduction Difficulty of formulating user queries –Insufficient knowledge of the collection –Insufficient.
A Markov Random Field Model for Term Dependencies Donald Metzler and W. Bruce Croft University of Massachusetts, Amherst Center for Intelligent Information.
INFO 624 Week 3 Retrieval System Evaluation
Learning to Advertise. Introduction Advertising on the Internet = $$$ –Especially search advertising and web page advertising Problem: –Selecting ads.
Ranking by Odds Ratio A Probability Model Approach let be a Boolean random variable: document d is relevant to query q otherwise Consider document d as.
Scalable Text Mining with Sparse Generative Models
ENDANGERED SPECIES AGST 3000 Agriculture, Society and the Natural World.
JOURNAL OF INFORMATION SCIENCE AND ENGINEERING 30, (2014) BERLIN CHEN, YI-WEN CHEN, KUAN-YU CHEN, HSIN-MIN WANG2 AND KUEN-TYNG YU Department of Computer.
CS344: Introduction to Artificial Intelligence Vishal Vachhani M.Tech, CSE Lecture 34-35: CLIR and Ranking in IR.
Probabilistic Model for Definitional Question Answering Kyoung-Soo Han, Young-In Song, and Hae-Chang Rim Korea University SIGIR 2006.
Minimal Test Collections for Retrieval Evaluation B. Carterette, J. Allan, R. Sitaraman University of Massachusetts Amherst SIGIR2006.
Philosophy of IR Evaluation Ellen Voorhees. NIST Evaluation: How well does system meet information need? System evaluation: how good are document rankings?
1 Cross-Lingual Query Suggestion Using Query Logs of Different Languages SIGIR 07.
1 Formal Models for Expert Finding on DBLP Bibliography Data Presented by: Hongbo Deng Co-worked with: Irwin King and Michael R. Lyu Department of Computer.
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
1 A Unified Relevance Model for Opinion Retrieval (CIKM 09’) Xuanjing Huang, W. Bruce Croft Date: 2010/02/08 Speaker: Yu-Wen, Hsu.
Query Routing in Peer-to-Peer Web Search Engine Speaker: Pavel Serdyukov Supervisors: Gerhard Weikum Christian Zimmer Matthias Bender International Max.
A Markov Random Field Model for Term Dependencies Donald Metzler W. Bruce Croft Present by Chia-Hao Lee.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
When Experts Agree: Using Non-Affiliated Experts To Rank Popular Topics Meital Aizen.
Thanks to Bill Arms, Marti Hearst Documents. Last time Size of information –Continues to grow IR an old field, goes back to the ‘40s IR iterative process.
Modern Information Retrieval: A Brief Overview By Amit Singhal Ranjan Dash.
Estimating Topical Context by Diverging from External Resources SIGIR’13, July 28–August 1, 2013, Dublin, Ireland. Presenter: SHIH, KAI WUN Romain Deveaud.
A Probabilistic Graphical Model for Joint Answer Ranking in Question Answering Jeongwoo Ko, Luo Si, Eric Nyberg (SIGIR ’ 07) Speaker: Cho, Chin Wei Advisor:
Mining the Web to Create Minority Language Corpora Rayid Ghani Accenture Technology Labs - Research Rosie Jones Carnegie Mellon University Dunja Mladenic.
Probabilistic Query Expansion Using Query Logs Hang Cui Tianjin University, China Ji-Rong Wen Microsoft Research Asia, China Jian-Yun Nie University of.
INTERESTING NUGGETS AND THEIR IMPACT ON DEFINITIONAL QUESTION ANSWERING Kian-Wei Kor, Tat-Seng Chua Department of Computer Science School of Computing.
Predicting Question Quality Bruce Croft and Stephen Cronen-Townsend University of Massachusetts Amherst.
Ranking in Information Retrieval Systems Prepared by: Mariam John CSE /23/2006.
Relevance Feedback Hongning Wang What we have learned so far Information Retrieval User results Query Rep Doc Rep (Index) Ranker.
Web Image Retrieval Re-Ranking with Relevance Model Wei-Hao Lin, Rong Jin, Alexander Hauptmann Language Technologies Institute School of Computer Science.
Empowering users to access information in the Digital Library Corin Anderson University of Washington.
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
Query Suggestion Naama Kraus Slides are based on the papers: Baeza-Yates, Hurtado, Mendoza, Improving search engines by query clustering Boldi, Bonchi,
Fish and Wildlife : Regulatory Framework and Challenges Cherise M. Oram STOEL RIVES LLP Hydrovision 2008 Ocean/Tidal/Stream Power Track 7D “Environmental.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
Personalization with user’s local data Personalizing Search via Automated Analysis of Interests and Activities 1 Sungjick Lee Department of Electrical.
21/11/20151Gianluca Demartini Ranking Clusters for Web Search Gianluca Demartini Paul–Alexandru Chirita Ingo Brunkhorst Wolfgang Nejdl L3S Info Lunch Hannover,
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Probabilistic Latent Query Analysis for Combining Multiple Retrieval Sources Rong Yan Alexander G. Hauptmann School of Computer Science Carnegie Mellon.
Finding Experts Using Social Network Analysis 2007 IEEE/WIC/ACM International Conference on Web Intelligence Yupeng Fu, Rongjing Xiang, Yong Wang, Min.
Language Model in Turkish IR Melih Kandemir F. Melih Özbekoğlu Can Şardan Ömer S. Uğurlu.
UWMS Data Mining Workshop Content Analysis: Automated Summarizing Prof. Marti Hearst SIMS 202, Lecture 16.
Date: 2012/08/21 Source: Zhong Zeng, Zhifeng Bao, Tok Wang Ling, Mong Li Lee (KEYS’12) Speaker: Er-Gang Liu Advisor: Dr. Jia-ling Koh 1.
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
NTNU Speech Lab Dirichlet Mixtures for Query Estimation in Information Retrieval Mark D. Smucker, David Kulp, James Allan Center for Intelligent Information.
Post-Ranking query suggestion by diversifying search Chao Wang.
Aruna Balasubramanian, Yun Zhou, W Bruce Croft, Brian N Levine and Arun Venkataramani Department of Computer Science, University of Massachusetts, Amherst.
26/01/20161Gianluca Demartini Ranking Categories for Faceted Search Gianluca Demartini L3S Research Seminars Hannover, 09 June 2006.
Automatic Labeling of Multinomial Topic Models
Divided Pretreatment to Targets and Intentions for Query Recommendation Reporter: Yangyang Kang /23.
1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis.
CS791 - Technologies of Google Spring A Web­based Kernel Function for Measuring the Similarity of Short Text Snippets By Mehran Sahami, Timothy.
Introduction to Information Retrieval Introduction to Information Retrieval Lecture 15: Text Classification & Naive Bayes 1.
Using Statistical Decision Theory and Relevance Models for Query-Performance Prediction Anna Shtok and Oren Kurland and David Carmel SIGIR 2010 Hao-Chin.
Biodiversity.
Evaluation of IR Systems
Information Retrieval
“I speak for the trees!”.
Authors: Wai Lam and Kon Fan Low Announcer: Kyu-Baek Hwang
Panagiotis G. Ipeirotis Luis Gravano
Presentation transcript:

Hierarchical Summaries By: Dawn J. Lawrie University of Massachusetts, Amherst for Search

Dawn J. Lawrie University of Massachusetts, Amherst 2 The Problem

Dawn J. Lawrie University of Massachusetts, Amherst 3 Possible Solution

Dawn J. Lawrie University of Massachusetts, Amherst 4 Possible Solution

Dawn J. Lawrie University of Massachusetts, Amherst 5 Solution: Automatic Hierarchies

Dawn J. Lawrie University of Massachusetts, Amherst 6 Strengths of Automatic Hierarchies Word-based summary Focus on topics of the documents Allows users to navigate through the results Easy to understand Bonus: Useful for summarizing documents

Dawn J. Lawrie University of Massachusetts, Amherst 7 Endangered Animals (29 10 ) marine mammals (18 8 ) Hand-generated hierarchy of 50 documents Query: “Endangered Species (Mammals)” Example legislation (6 4 )permits (10 2 )Critical Habitat (16 0 ) Endangered plants (7 0 ) Ecosystem Management (2 0 ) Threatened (1 0 ) Endangered Species Act (1 0 ) mammals (17 10 ) fish (7 0 ) birds (3 0 ) insects (3 0 ) amphibians (1 0 ) sea lions (2 2 ) manatees (1 1 ) whales (7 4 ) jaguars (2 0 ) marine (12 8 ) deer (1 1 ) habitat protection (1 1 ) rats (1 0 ) Hawaii (3 0 ) California (2 0 ) Utah (1 0 ) Virginia (1 0 ) Melicope Species (1 0 ) Wainae Plant Cluster Recovery Plan (1 0 ) Waianae Mountains (1 0 )

Dawn J. Lawrie University of Massachusetts, Amherst 8 Proposed Framework Document Set Language Model Term Selection Algorithm Hierarchy “Term” = word or phrase

Dawn J. Lawrie University of Massachusetts, Amherst 9 Challenges Selecting terms for the hierarchy Displaying the hierarchy Showing that it works

Dawn J. Lawrie University of Massachusetts, Amherst 10 Outline Introduction Description of framework for creating hierarchies Examples Methods of evaluation Future Improvements

Dawn J. Lawrie University of Massachusetts, Amherst 11 Methodology Build probabilistic word model of documents Find “best” terms On topic Predictive Recursive definition creates hierarchy

Dawn J. Lawrie University of Massachusetts, Amherst 12 Term characteristics Why topicality? Distinguish topic terms from the rest of the vocabulary The Secretary of Interior listed bald eagles south of the 40 th parallel as endangered under the Endangered Species Preservation Act of Why predictiveness? Topic words can be strongly related Represent different facets of the vocabulary Example: P(“Endangered”|”Stellar sea lions”) = 1.00

Dawn J. Lawrie University of Massachusetts, Amherst 13 Statistical Model A T refers to topicality with respect to topic T Find if the word w is in set T B refers to predictiveness Precondition for other terms to occur Find if word w is in set P

Dawn J. Lawrie University of Massachusetts, Amherst 14 Probabilistic Word Model Captures statistical information about text Called a “language model” in speech recognition Provides basis for estimation of probabilities

Dawn J. Lawrie University of Massachusetts, Amherst 15 Estimating Topicality Use term’s contribution to relative entropy Compares two models using K-L divergence Model of documents in hierarchy Model of general English

Dawn J. Lawrie University of Massachusetts, Amherst 16 KL Example mammal fishery species marine endangered

Dawn J. Lawrie University of Massachusetts, Amherst 17 Estimating Predictiveness Relates the vocabulary to a set of candidate topic terms Use conditional probability - P x ( t | v ) x is the maximum distance between t and v mammal marine mammal marine species fishery species fishery t v P(t|v)P(t|v)

Dawn J. Lawrie University of Massachusetts, Amherst 18 Dominating Set Approximation Interpret predictive language model as graph edges weighted by the conditional probability Finds terms that are connected to lots of terms with a high weight Chooses topic terms until vocabulary is dominated (predicted)

Dawn J. Lawrie University of Massachusetts, Amherst 19 Term Selection Example P(t|v) v t

Dawn J. Lawrie University of Massachusetts, Amherst 20 Generating a Summary 4-step process (1) Preprocess document set (2) Generate a language model (3) Select the terms (4) Create a Hierarchy recursive

Dawn J. Lawrie University of Massachusetts, Amherst 21 Outline Introduction Description of framework for creating hierarchies Examples Methods of evaluation Future Improvements

Dawn J. Lawrie University of Massachusetts, Amherst 22 Example Hierarchies Generated from 50 documents retrieved for the query: Endangered Species - Mammals Demonstrate the difference between using different topic models Web hierarchy using same query

Dawn J. Lawrie University of Massachusetts, Amherst 23 endangered (8 6 ) Act (4 1 ) State (3 2 ) Committee (4 3 ) address (8 5 ) operations (4 3 ) incidental take (4 2 ) NMFS (6 4 ) population (3 2 ) commercial fishing operations (4 2 ) regulations (12 4 ) fish (11 7 ) permit (14 6 ) number (9 3 ) bill (5 1 ) Secretary (7 3 ) research (10 5 ) amended (15 4 ) Uniform Topic Model Hierarchy marine (18 7 ) species (43 9 ) plan (19 2 ) marine mammals (18 7 )

Dawn J. Lawrie University of Massachusetts, Amherst 24 KL-Topic Model Hierarchy species (43 9 ) Marine Mammal Protection Act (7 3 ) marine mammals (18 7 ) management plan (5 1 ) marine (18 7 ) Endangered Species Act (29 4 ) endangered species (20 4 ) habitat (28 3 ) mammals (12 6 ) Marine Mammal Commission (2 1 ) fish (27 7 ) National Marine Fisheries Service (11 3 ) Act (31 3 ) permit (16 4 ) protection (24 4 ) marine mammal stocks (2 0 ) marine mammal species (4 2 ) fishery (5 3 ) Secretary (4 2 ) NMFS (8 3 ) stock (5 1 ) fish species (3 2 ) MMPA (5 1 ) incidental (7 4 ) research (6 3 )

Dawn J. Lawrie University of Massachusetts, Amherst 25 Query-Topic Model Hierarchy marine mammal species (4 2 ) incidental (7 4 ) fish species (3 2 ) threatened species (1 0 ) requirements (8 4 ) population (6 2 ) MMPA (5 1 ) Protected Resources (2 1 ) marine mammal stocks (2 0 ) animals (4 2 ) habitat (28 3 ) endangered species (20 4 ) mammals (12 6 ) listed species (11 0 ) endangered species program (3 0 ) Act (31 3 ) National Marine Fisheries Service (11 3 ) Marine Mammal Protection Act (7 3 ) Wildlife Service (17 2 ) listed (26 0 ) Endangered Species Act (33 5 ) marine mammals (18 7 ) Federal (23 2 ) marine (18 7 ) species (43 9 )

Dawn J. Lawrie University of Massachusetts, Amherst 26 Web Hierarchies Submit query to a web search engine Gather titles and snippets of documents Text considered a document Documents are about 30 words

Dawn J. Lawrie University of Massachusetts, Amherst 27 species of marine mammals (1) Listed Species (1) Species Information (1) Endangered Species Act (8) Protected Resources (2) sea otter (2) whales (13) dolphins (7) Cetaceans (2) marine (76) Mammals species (4) Canadian Endangered Species (3) federal Endangered Species (1) marine mammals (91) Endangered Mammals (22) threatened (144) species of mammals (27) endangered mammal species (4) birds (140) British mammals (4) animal species (1) Critically Endangered Mammals (2) Animal Info (2) Ecosystems (2) Scientists (2) species of marine mammals (1) Endangered Species Coalition (2) Endangered Spaces (2) List of Endangered Species (5) marine mammals (97) birds (114) Endangered Mammals (13) threatened (78) small mammals (13) large mammals (12) Example of Web Hierarchy Endangered Species (440) endangered (491) mammals (600) terrestrial mammals (2) endangered marine species (2) Species Management (2) marine species (4) listing of species (1) protected species (2) native species (1) Candidate species (2) 100 species (1) new species (1)

Dawn J. Lawrie University of Massachusetts, Amherst 28 Outline Introduction Description of framework for creating hierarchies Examples Methods of evaluation Future Improvements

Dawn J. Lawrie University of Massachusetts, Amherst 29 Evaluations Summary Evaluation Tests how well the topic terms chosen predict the vocabulary Access Evaluation Compare number of documents a user can find Relevance Evaluation Path length to find all relevant documents

Dawn J. Lawrie University of Massachusetts, Amherst 30 Automatic Evaluation Test Set Use 50 standard queries Document sets 500 documents retrieved from TREC volumes 4 and 5 (have relevance judgments) 200 documents retrieved from a news database 1000 titles and snippets retrieved using Google™ Search Engine

Dawn J. Lawrie University of Massachusetts, Amherst 31 Evaluating Hypotheses Use KL-topic model Use sub-collections Summary Access Relevance TREC Collection and News Documents -Denotes an evaluation confirmed hypothesis -Denotes evaluation showed no significant difference

Dawn J. Lawrie University of Massachusetts, Amherst 32 Web Document Evaluation Results completely different Best hierarchy uniform topic model Hierarchies do not look as good to human inspection

Dawn J. Lawrie University of Massachusetts, Amherst 33 User Study Include 12 to 16 users Compare ranked list and hierarchy to ranked list alone Users asked to find all instances that are relevant to the query Only have to identify one document about a particular instance Study includes 10 queries

Dawn J. Lawrie University of Massachusetts, Amherst 34 Future Work Complete user study Failure Analysis Explore the use of topic hierarchies in other organizational tasks Personal collections of documents s

Dawn J. Lawrie University of Massachusetts, Amherst 35 Conclusions Developed a formal framework for topic hierarchies Created hierarchies from full text and snippets of documents Verified intuition concerning hierarchies generated from full text

Dawn J. Lawrie University of Massachusetts, Amherst 36 Questions? Demo:

Dawn J. Lawrie University of Massachusetts, Amherst 37 Improving Topicality Estimate Estimate topicality using a query model Emphasizes query related terms Improve model of English with sub-collections Distinguishes between terms that are frequently used in a genre and topic terms

Dawn J. Lawrie University of Massachusetts, Amherst 38 Key Ideas Language models are created from documents Topicality and predictiveness are used to choose terms Topicality is estimated using Kullback-Leibler divergence Predictiveness is estimated by calculating conditional probabilities

Dawn J. Lawrie University of Massachusetts, Amherst 39 Key Ideas Showed the effect of using topic model Observed trade-off between snippets and full text of documents

Dawn J. Lawrie University of Massachusetts, Amherst 40 Summary Evaluation Expected Mutual Information Measure Two sets deviates from stochastic independence Shows how well the topic terms chosen predict the vocabulary

Dawn J. Lawrie University of Massachusetts, Amherst 41 Hierarchy Terms vs. Top TF.IDF Terms TF.IDF popular term weight Commonly used as a method of naming clusters Compare equal number of unique terms in hierarchy to top TF.IDF Terms Results Hierarchies always significantly better at summarizing documents

Dawn J. Lawrie University of Massachusetts, Amherst 42 Compare number of documents that are accessible Example policy Examine parts of the hierarchy with 20 or fewer documents Look at the top 200 documents in a ranked list Access Evaluation

Dawn J. Lawrie University of Massachusetts, Amherst 43 Level Size = 5 Rank 400 Rank 350 Rank 300 Rank 250 Rank 200 Hier. Topics = 50 Hier. Topics = 45 Hier. Topics = 40 Rank 150 Hier. Topics = 35 Hier. Topics = 30 Hier. Topics = 25 Hier. Topics = 20 Hier. Topics = 15 Hier. Topics = 10 Rank 100 Hier. Topics = 5 Rank 50 Level Size = 10 Rank 500 Rank 450 Rank 400 Rank 350 Rank 300 Hier. Topics = 50 Hier. Topics = 45 Hier. Topics = 40 Hier. Topics = 35 Rank 250 Hier. Topics = 30 Hier. Topics = 25 Hier. Topics = 20 Hier. Topics = 15 Hier. Topics = 10 Rank 200 Hier. Topics = 5 Rank 150 Level Size = 20 Rank 500 Rank 450 Hier. Topic = 50 Hier. Topic = 45 Hier. Topic = 40 Rank 400 Hier. Topics = 35 Hier. Topics = 30 Hier. Topics = 25 Hier. Topics = 20 Hier. Topics = 15 Hier. Topics = 10 Rank 350 Hier. Topics = 5 Rank 300 Rank 250 Rank 200 Rank 150 Level Size = 15 Rank 500 Rank 450 Rank 400 Rank 350 Hier. Topics = 50 Hier. Topics = 45 Hier. Topics = 40 Hier. Topics = 35 Hier. Topics = 30 Hier. Topics = 25 Hier. Topics = 20 Hier. Topics = 15 Rank 300 Hier. Topics = 10 Hier. Topics = 5 Rank 250 Rank 200 Rank 150 Hierarchy vs. Ranked List

Dawn J. Lawrie University of Massachusetts, Amherst 44 Relevance Evaluation Calculate average path to a relevant document Assumptions one does not read extraneous menus one reads all documents at a node Ignores relevant documents that are not in the hierarchy Smaller score denotes a better hierarchy

Dawn J. Lawrie University of Massachusetts, Amherst 45 Key Ideas Automatic evaluations have confirmed hypotheses for hierarchies created from full text of documents User study is necessary for determining how well people can use hierarchies