Web Search Engines –
Browsing Services Search Engine Services Web Pages Bag of Words Two semantics extremes Two service extremes ???
–
Web ……
version: 1.0// version number url: URL origin: original URL date: Tue, 15 Apr :13:06 GMT // time of harvest ip: // IP address unzip-length: // If included, the data must be compressed length: 18133// data length // a blank line XXXXXXXX// the followings are data part XXXXXXXX …. XXXXXXXX// data end // insert a new line
File Organizations (Indexes) Choices for accessing data during query evaluation Scan the entire collection –Typical in early (batch) retrieval systems –Computational and I/O costs are O(characters in collection) –Practical for only small text collections –Large memory systems make scanning feasible Use indexes for direct access –Evaluation time O(query term occurrences in collection) –Practical for large collections –Many opportunities for optimization Hybrids: Use small index, then scan a subset of the collection
Indexes What should the index contain? Database systems index primary and secondarykeys –This is the hybrid approach –Index provides fast access to a subset of database records –Scan subset to find solution set IR Problem: Cannot predict keys that people will use in queries –Every word in a document is a potential search term IR Solution: Index by all keys (words) full text indexes
Index Contents The contents depend upon the retrieval model Feature presence/absence –Boolean –Statistical (tf, df, ctf, doclen, maxtf) –Often about 10% the size of the raw data, compressed Positional –Feature location within document –Granularities include word, sentence, paragraph, etc –Coarse granularities are less precise, but take less space –Word-level granularity about 20-30% the size of the raw data,compressed
Indexes: Implementation Common implementations of indexes –Bitmaps –Signature files –Inverted files Common index components –Dictionary (lexicon) –Postings document ids word positions No positional data indexed
Inverted Files
Word-Level Inverted File
Inverted Search Algorithm 1.Find query elements (terms) in the lexicon 2.Retrieve postings for each lexicon entry 3.Manipulate postings according to the retrieval model
Word-Level Inverted File Query: 1.porridge & pot (BOOL) 2.porridge pot (BOOL) 3. porridge pot (VSM)VSM lexicon posting Answer
A Brief history of Modern Information Retrieval In 1945, Vannevar Bush published "As We May Think" in the Atlantic monthly. In the 1960s, the SMART system by Gerard Salton and his students Cranfield evaluations done by Cyril Cleverdon The 1970s and 1980s saw many developments built on the advances of the 1960s. In 1992 with the inception of Text Retrieval Conference. The algorithms developed The algorithms developed in IR were employed for searching the Web from 1996.
Clustering of SIGIR papers by topic vs. year
Question answering
Clustering
Inverted files & Implementations
Message understanding & TDT
Filtering
Hypertext IR, Multiple evidence
Probabilistic & Language models
Distributed IR
Evaluation
Topic distillation & Linkage retrieval
Text categorisation
Document summarisation
Cross lingual
CIIR, University of Massachusetts LTI, Carnegie Mellon University The Stanford University DB Group Microsoft Research Asia TREC,,
Lemur
Lemur Toolkit LM IR research system –ad hoc, distributed retrieval, cross-language IR, summarization, filtering, and classification : – – Simple Language Model – Language Model : –C and C++ –Unix / Windows –Current Version 3.1
MRA: Towards Next Generation Web Search From Pages to Blocks –Analyze the Web at finer granularity From Surface Web to Deep Web –Unleash the huge assets of high-value information From Unstructure to Structure –Provide well organized results From relevance to intelligence –Contribute knowledge discovery with search From Desktop Search to Mobile Search –Bridge physical world search to digital world search
The Stanford Univ. DB Group WebBase –Crawling, storage, indexing, and querying of large collections of Web pages. Digital Libraries –Infrastructure and services for creating, disseminating, sharing and managing information
TREC Conference Established in 1992 to evaluate large-scale IR –Retrieving documents from a gigabyte collection Has run continuously since then –TREC 2004(13 th ) meeting is in November Run by NISTs Information Access Division Probably most well known IR evaluation setting –Started with 25 participating organizations in 1992 evaluation –In 2003, there were 93 groups from 22 different countries Proceedings available on-line ( ) –Overview of TREC 2003 at f
TREC consists of IR research tracks –Ad hoc, routing, confusion ( scanned documents, speech recognition ), video, filtering, multilingual ( cross-language, Spanish, Chinese ), question answering, novelty, high precision, interactive, Web, database merging, NLP, … Each track works on roughly the same model –November: track approved by TREC community –Winter: tracks members finalize format for track –Spring: researchers train system based on specification –Summer: researchers carry out format evaluation Usually a blind evaluation: research do not know answer –Fall: NIST carries out evaluation –November: Group meeting (TREC) to find out: How well your site did How others tackled the program –Many tracks are run by volunteers outside of NIST (e.g. Web) Coopetition model of evaluation –Successful approaches generally adopted in next cycle TREC General Format
TREC Tracks
Summary of VLC/Web Track evaluation
Tianwang
CWT100g !
/8.8 = 28.4%
TEAMNAME TD- RUNS NPHP- RUNS APEX 55 ANS32 TRS 52 MUMIAN131 MUMIAN221 SCUTDB55 WLL 1 pooling google,yisou,baidu,sogou,zhongsou SE
TIANWANG_RUN
( :
Vector Space Model d q m TFIDF : ( tf,idf ) BACK
Query Answer 1.porridge & pot (BOOL) –d2 2.porridge pot (BOOL) –null 3. porridge pot (VSM) –d2 > d1>d5 –Next page BACK
CIIR- Center for Intelligent Information One of the leading research groups in IR –improving the probabilistic models, –first description of a retrieval system based on statistical language models. –introduced and improved a number of techniques for text and query representation –automatically representing databases and combining local searches for DIR –first high capacity probabilistic filtering architecture –define and evaluate the first versions of event detection and tracking software –earliest research on ranking and representation techniques for Asian languages –first approaches to information extraction that emphasized learning –novel techniques for indexing images and video
CIIR cont. Research –more than 500 journal and refereed conference papers over the past 12 years (52 submissions in 2003). industrial and government collaboration –INQUERY –licensed our software to nearly 300 sites Education –20 Ph.D.s, 29 M.S. –123/145, 34/4 graduate/undergraduate
CIIR cont. Personnel –Faculty4(W. BRUCE CROFT) –Technical personel10 –Graduate student34/10 Groups –IESL:Information Extraction and Synthesis Laboratory –IR :Information Retrieval Laboratory –MIR :Multimedia Indexing and Retrieval Laboratory The CIIR is currently concentrating on the unsolved long- term research problems that underlie effective information retrieval –text representation, –query acquisition, –retrieval models
LTI : Language Technologies Machine Translation, Natural Language Processing, Speech, and Information Retrieval IR Projects (Jamie Callan and Yiming Yang ) –Adaptive Information Filtering –Distributed Information Retrieval / Federated Search – Classification and Prioritization –Minerva: Web Mining for Question Answering –MuchMore: Translingual Information Retrieval –JAVELIN: Open-Domain Question Answering BACK