Using Information Content to Evaluate Semantic Similarity in a Taxonomy Presenter: Cosmin Adrian Bejan Philip Resnik Sun Microsystems Laboratories.

Slides:



Advertisements
Similar presentations
Date: 2014/05/06 Author: Michael Schuhmacher, Simon Paolo Ponzetto Source: WSDM’14 Advisor: Jia-ling Koh Speaker: Chen-Yu Huang Knowledge-based Graph Document.
Advertisements

Ciro Cattuto, Dominik Benz, Andreas Hotho, Gerd Stumme Presented by Smitashree Choudhury.
Linked data: P redicting missing properties Klemen Simonic, Jan Rupnik, Primoz Skraba {klemen.simonic, jan.rupnik,
Semantic Similarity in a Taxonomy By Ankit Ramteke ( ) Bibek Behera ( ) Karan Chawla ( )
A UTOMATICALLY A CQUIRING A S EMANTIC N ETWORK OF R ELATED C ONCEPTS Date: 2011/11/14 Source: Sean Szumlanski et. al (CIKM’10) Advisor: Jia-ling, Koh Speaker:
Variance reduction techniques. 2 Introduction Simulation models should be coded such that they are efficient. Efficiency in terms of programming ensures.
MSS 905 Methods of Missiological Research
Introduction to Information Retrieval (Manning, Raghavan, Schutze) Chapter 6 Scoring term weighting and the vector space model.
Topic 6: Introduction to Hypothesis Testing
Evaluating Search Engine
Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP February 23, 2011.
Measures of Distributional Similarity Presenter: Cosmin Adrian Bejan Lillian Lee Department of Computer Science Cornell University.
Semantic text features from small world graphs Jure Leskovec, IJS + CMU John Shawe-Taylor, Southampton.
Semantic distance in WordNet: An experimental, application-oriented evaluation of five measures Presenter: Cosmin Adrian Bejan Alexander Budanitsky and.
Frequency Estimates for Statistical Word Similarity Measures Presenter: Cosmin Adrian Bejan Egidio Terra and C.L.A. Clarke School of Computer Science University.
Computing Trust in Social Networks
June 19-21, 2006WMS'06, Chania, Crete1 Design and Evaluation of Semantic Similarity Measures for Concepts Stemming from the Same or Different Ontologies.
Taming the Underlying Challenges of Reliable Multihop Routing in Sensor Networks.
Designing clustering methods for ontology building: The Mo’K workbench Authors: Gilles Bisson, Claire Nédellec and Dolores Cañamero Presenter: Ovidiu Fortu.
Learning syntactic patterns for automatic hypernym discovery Rion Snow, Daniel Jurafsky and Andrew Y. Ng Prepared by Ang Sun
Distributional clustering of English words Authors: Fernando Pereira, Naftali Tishby, Lillian Lee Presenter: Marian Olteanu.
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Alignment Inclusive Assessment Seminar Brian Gong Claudia.
On Self Adaptive Routing in Dynamic Environments -- A probabilistic routing scheme Haiyong Xie, Lili Qiu, Yang Richard Yang and Yin Yale, MR and.
Albert Gatt LIN 3098 Corpus Linguistics. In this lecture Some more on corpora and grammar Construction Grammar as a theoretical framework Collostructional.
SI485i : NLP Set 10 Lexical Relations slides adapted from Dan Jurafsky and Bill MacCartney.
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
CS324e - Elements of Graphics and Visualization Java Intro / Review.
Personalisation Seminar on Unlocking the Secrets of the Past: Text Mining for Historical Documents Sven Steudter.
Copyright © Curt Hill Other Trees Applications of the Tree Structure.
A Compositional Context Sensitive Multi-document Summarizer: Exploring the Factors That Influence Summarization Ani Nenkova, Stanford University Lucy Vanderwende,
Clustering User Queries of a Search Engine Ji-Rong Wen, Jian-YunNie & Hon-Jian Zhang.
When Experts Agree: Using Non-Affiliated Experts To Rank Popular Topics Meital Aizen.
1 Statistical NLP: Lecture 9 Word Sense Disambiguation.
WORD SENSE DISAMBIGUATION STUDY ON WORD NET ONTOLOGY Akilan Velmurugan Computer Networks – CS 790G.
Efficient Deployment Algorithms for Prolonging Network Lifetime and Ensuring Coverage in Wireless Sensor Networks Yong-hwan Kim Korea.
10/22/2015ACM WIDM'20051 Semantic Similarity Methods in WordNet and Their Application to Information Retrieval on the Web Giannis Varelas Epimenidis Voutsakis.
Introduction to Probability  Probability is a numerical measure of the likelihood that an event will occur.  Probability values are always assigned on.
Selfishness, Altruism and Message Spreading in Mobile Social Networks September 2012 In-Seok Kang
Modelling Human Thematic Fit Judgments IGK Colloquium 3/2/2005 Ulrike Padó.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
Huffman coding Content 1 Encoding and decoding messages Fixed-length coding Variable-length coding 2 Huffman coding.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
SRL: A Bidirectional Abstraction for Unidirectional Ad Hoc Networks. Venugopalan Ramasubramanian Ranveer Chandra Daniel Mosse.
Iterative Translation Disambiguation for Cross Language Information Retrieval Christof Monz and Bonnie J. Dorr Institute for Advanced Computer Studies.
Algorithmic Detection of Semantic Similarity WWW 2005.
Authors: Marius Pasca and Benjamin Van Durme Presented by Bonan Min Weakly-Supervised Acquisition of Open- Domain Classes and Class Attributes from Web.
Class 2 Probability Theory Discrete Random Variables Expectations.
Ngram models and the Sparcity problem. The task Find a probability distribution for the current word in a text (utterance, etc.), given what the last.
Finding Experts Using Social Network Analysis 2007 IEEE/WIC/ACM International Conference on Web Intelligence Yupeng Fu, Rongjing Xiang, Yong Wang, Min.
1 CSC 594 Topics in AI – Text Mining and Analytics Fall 2015/16 3. Word Association.
Using Semantic Relatedness for Word Sense Disambiguation
Graphs & Matrices Todd Cromedy & Bruce Nicometo March 30, 2004.
A Knowledge-Based Search Engine Powered by Wikipedia David Milne, Ian H. Witten, David M. Nichols (CIKM 2007)
Comparing Word Relatedness Measures Based on Google n-grams Aminul ISLAM, Evangelos MILIOS, Vlado KEŠELJ Faculty of Computer Science Dalhousie University,
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
Acquisition of Categorized Named Entities for Web Search Marius Pasca Google Inc. from Conference on Information and Knowledge Management (CIKM) ’04.
Investigating semantic similarity measures across the Gene Ontology: the relationship between sequence and annotation Bioinformatics, July 2003 P.W.Load,
NLP.
2/10/2016Semantic Similarity1 Semantic Similarity Methods in WordNet and Their Application to Information Retrieval on the Web Giannis Varelas Epimenidis.
Semantic Grounding of Tag Relatedness in Social Bookmarking Systems Ciro Cattuto, Dominik Benz, Andreas Hotho, Gerd Stumme ISWC 2008 Hyewon Lim January.
Copyright © Curt Hill Other Trees Applications of the Tree Structure.
Copyright ©2004 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 4-1 Probability and Counting Rules CHAPTER 4.
WordNet::Similarity Measuring the Relatedness of Concepts Yue Wang Department of Computer Science.
From NARS to a Thinking Machine Pei Wang Temple University.
An Adaptive Learning with an Application to Chinese Homophone Disambiguation from Yue-shi Lee International Journal of Computer Processing of Oriental.
Statistical NLP: Lecture 9
An Empirical Study of Property Collocation on Large Scale of Knowledge Base 龚赛赛
Giannis Varelas Epimenidis Voutsakis Paraskevi Raftopoulou
Unsupervised Word Sense Disambiguation Using Lesk algorithm
Statistical NLP : Lecture 9 Word Sense Disambiguation
Presentation transcript:

Using Information Content to Evaluate Semantic Similarity in a Taxonomy Presenter: Cosmin Adrian Bejan Philip Resnik Sun Microsystems Laboratories

2 Overview  The paper presents an alternative way to evaluate semantic similarity in a taxonomy, based on the notion of information content.  It provides a way of adapting a static knowledge structure to multiple contexts by combining a taxonomic structure with empirical probability estimates.

3 Introduction  semantic similarity is a special case of semantic relatedness:  for example, cars - gasoline would seem to be more closely related than, say, cars - bicycles, but the latter pair are certainly more similar.  a natural way to evaluate semantic similarity in a taxonomy is to evaluate the distance between the nodes corresponding to the items being compared – the shorter the path from one node to another, the more similar they are.  but one of the problems with this approach is that it relies on the notion that links in the taxonomy represent uniform distances. Examples:  rabbit ears IS-A television antenna  phytoplankton IS-A living thing

4 Similarity of two concepts  Intuitively, one key in the similarity of two concepts is the extent to which they share information in common, indicated in an IS-A taxonomy by a highly specific concept that subsumes them both.  By associating probabilities with concepts in the taxonomy, it is possible to capture the same idea, but avoiding the unreliability of edge distances

5 Similarity and Information Content  Let:  C – the set of concepts in an IS-A taxonomy, permitting multiple inheritance  Function p:C->[0,1], such that for any c  C, p(c) is the probability of encountering an instance of concept c.  Notes:  p is monotonic as one moves up the taxonomy: if c 1 IS-A c 2, then p(c 1 ) ≤ p(c 2 )  if the taxonomy has a unique top node then its probability is 1.

6 Similarity and Information Content  the information content of a concept c can be quantified as negative the log likelihood -log p(c)  as probability increases, informativeness decreases, so the more abstract a concept, the lower its information content.  the information shared by two concepts is indicated by the information content of the concepts that subsumes them in the taxonomy: where S(c 1, c 2 ) is the set of concepts that subsume both c 1 and c 2

7 Similarity and Information Content  although similarity is computed by considering all upper bounds for the two concepts, the information measure has the effect of identifying minimal upper bounds since no class is less informative than its superordinates.

8 Implementation  WordNet – the taxonomy of concepts and compound nominals  frequencies of concepts in the taxonomy were estimated using noun frequencies from Brown Corpus of American English  Each noun that occurred in the corpus was counted as an occurrence of each taxonomic class containing it: where words(c) is the set of words subsumeb by concept c  concept probabilities were computed simply as relative frequency: where N is the total number of nouns observed (and exist in WN)

9 Task  one way to evaluate computational measures of semantic similarity is to find a correlation with human similarity ratings.  replicated experiment with human subjects: giving ten subjects 30 noun pairs they were asked to rate “similarity of meaning” for each pair on a scale from 0 (no similarity) to 4 (perfect synonymy).  the average correlation over the 10 subjects was r=0.8848

10 Results  Three similarity measures were used:  Similarity measurement using information content  A variant on the second counting method  a measure that simply uses the probability of a concept, rather than the information content: