Presentation is loading. Please wait.

Presentation is loading. Please wait.

Semantic Similarity Methods in WordNet and their Application to Information Retrieval on the Web Yizhe Ge.

Similar presentations


Presentation on theme: "Semantic Similarity Methods in WordNet and their Application to Information Retrieval on the Web Yizhe Ge."— Presentation transcript:

1 Semantic Similarity Methods in WordNet and their Application to Information Retrieval on the Web
Yizhe Ge

2 Overview Backgrounds WordNet and Semantic Similarity Methods
Semantic Similarity Retrieval Model (SSRM) Web Retrieval System Experimental Results Conclusions Brief introductions of the backgrounds

3 Backgrounds Lack of common terms != not related
SSRM vs. Classical Models Semantically Similar vs. Lexicographic Term Matching Information retrieval is currently being widely used in domains from database systems to search engines. The main idea is to locate doc- uments that contain terms that the users specify in queries. classical information retrieval models (e.g., Vector Space, Probabilistic, Boolean) [14] is based on lexicographic term matching. However, the lack of common terms in two documents does not nec- essarily mean that the documents are not related. two terms can be semantically similar (e.g., can be synonyms or have similar meaning) although they are lexicographically different Therefore, classical retrieval methods will fail to retrieve documents with semantically similar terms. This is exactly the problem this paper is addressing

4 WordNet and Semantic Similarity Methods
An online lexical reference system Hierarchy and Relationships Hyponym/Hypernym (Is-A) Meronym/Holonym (Part-Of) WordNet is developed at the Princeton University It is an on-line lexical reference system, which attempts to model the lexical knowledge of the native English speakers It contains around 100,000 terms, including Nouns, verbs, adjectives and adverbs, which are organized into hierarchies

5 WordNet and Semantic Similarity Methods
Semantic Similarity Methods (4 main) Edge Counting Method Length of path Information Content Method Probability of occurrence Feature Based Method Properties Hybrid Method Edge Counting Methods: Measure the similarity be- tween two terms (concepts) as a function of the length of the path linking the terms Information Content Methods: Measure the difference in information content of the two terms as a function of their probability of occurrence in a corpus

6 Semantic Similarity Retrieval Model (SSRM)
Queries and Documents  Term Vectors Each term in vector is represented by its weight TFIDF is used for computing the weight Queries and documents are syntactically analyzed and re- duced into term (noun) vectors. A term is usually defined as a stemmed non stop-word. Very infrequent or very frequent terms are eliminated Each term in vector is represented by its weight We use the TFIDF model for computing the weight

7 Semantic Similarity Retrieval Model (SSRM)
Vector Space Model (VSM) Cosine Similarity Traditionally, for vector space models, the similarity between two documents is computed as the cosine of the inner product between their document vectors Where qi and di are the weights in the two vector representations Given a query, all documents are ranked according to their similarity with the query Once again, the lack of common terms in two documents does not necessarily mean that the documents may not contain the same terms. Semantically similar concepts maybe expressed in different words in the documents and the queries, and direct comparison by word-based VSM is not effective. For example, VSM will not recognize synonyms or semantically similar terms (e.g., ”car”, ”automobile”).

8 Semantic Similarity Retrieval Model (SSRM)
SSRM Works in Three Steps Term Re-weighting Term Expansion Document Similarity

9 Semantic Similarity Retrieval Model (SSRM)
Term Re-weighting Give higher weights to semantically similar query terms The weight qi of each query term i is adjusted based on its relationships with other se- mantically similar terms j within the same vector where t is a user defined threshold. This formulae suggests assigning higher weights to semantically similar terms within the query (e.g., ”railway”, ”train”, ”metro”). The weights of non-similar terms remain unchanged (e.g., ”train”, ”house”).

10 Semantic Similarity Retrieval Model (SSRM)
Term Expansion Hypo/hypernyms Each query term is assigned the weight as follows: First, the query is augmented by syn- onym terms (the most common sense is taken). Then, the query is augmented by hyponyms and hypernyms which are semantically similar to terms already in the query. Fig. 3 illustrates this process: Each query term is represented by its WordNet tree hierarchy. The neighborhood of the term is examined and all terms with similarity greater than threshold T (T = 0.9 in this work) are also included in the query vector. This expansion may include terms more than one level higher or lower than the original term. Then, each query term i is assigned a weight as follows where n is the number of hyponyms of each expanded term j. It is possible for a term to introduce terms that al- ready existed in the query. It is also possible that the same term is introduced by more than one other terms.

11 Semantic Similarity Retrieval Model (SSRM)
Document Similarity i and j are terms in the query and the document respectively In this calculation, query terms are expanded and re-weighted according to the previous steps while doc- ument terms dj are computed as tf · idf terms

12 Semantic Similarity Retrieval Model (SSRM)
Pro Relax the classical assumptions Con Quadratic time complexity: O(n^2) SSRM relaxes the requirement of classical retrieval models that conceptually similar terms are mutually independent It has a quadratic time complexity as opposed to the linear time complexity of VSM. Expanding and re-weighting is fast for queries (queries are short in most cases specifying only a few terms) but not for documents with many terms

13 Web Retrieval System Two inverted files associate terms with their intra and inter document frequencies and allow for fast computation of term vectors. These descriptions are syntactically analyzed and reduced into vectors of stemmed terms (nouns) which are matched against the queries.

14 Experimental Results SSRM vs. VSM
20 most frequent short image queries from Google 5 human referees Precision and Recall For the evaluations, 20 queries were selected from the list of the most frequent Google image queries These are rather short queries containing between 1 and 4 terms. The evaluation is based on human relevance judgments by 5 hu- man referees. Each referee evaluated a subset of 4 queries for both methods. Each method is represented by a precision-recall curve. Each query retrieves the best 50 answers and each point on a curve is the average precision and recall over 20 queries. Fig. 5.2 indicates that SSRM is far more effective than VSM achieving up to 30% better precision and up to 20% better recall. A closer look into the results reveals that the efficiency of SSRM is mostly due to the contribution of non-identical but semantically similar terms. VSM (like most classical retrieval models relying on lexicographic term matching) ignore this information.

15 Conclusions SSRM Can Work in Conjunction with Any Taxonomic Ontology
SSRM Performs Better Than VSM Future: extend SSRM to work with compounded terms SRM can work in conjunction with any taxonomic ontology (e.g., application specific ontolo- gies). The evaluation demonstrated very promising perfor- mance improvements over the Vector Space Model (VSM), which is the state-of-the-art document retrieval method.


Download ppt "Semantic Similarity Methods in WordNet and their Application to Information Retrieval on the Web Yizhe Ge."

Similar presentations


Ads by Google