2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Distant Supervision for Emotion Classification in Twitter posts 1/17.
Linear Model Incorporating Feature Ranking for Chinese Documents Readability Gang Sun, Zhiwei Jiang, Qing Gu and Daoxu Chen State Key Laboratory for Novel.
Playing the Telephone Game: Determining the Hierarchical Structure of Perspective and Speech Expressions Eric Breck and Claire Cardie Department of Computer.
Sentiment Analysis An Overview of Concepts and Selected Techniques.
D ETERMINING THE S ENTIMENT OF O PINIONS Presentation by Md Mustafizur Rahman (mr4xb) 1.
CS Word Sense Disambiguation. 2 Overview A problem for semantic attachment approaches: what happens when a given lexeme has multiple ‘meanings’?
Predicting the Semantic Orientation of Adjective Vasileios Hatzivassiloglou and Kathleen R. McKeown Presented By Yash Satsangi.
Predicting the Semantic Orientation of Adjectives
Learning Subjective Nouns using Extraction Pattern Bootstrapping Ellen Riloff, Janyce Wiebe, Theresa Wilson Presenter: Gabriel Nicolae.
Learning Subjective Adjectives from Corpora Janyce M. Wiebe Presenter: Gabriel Nicolae.
Sentence Classifier for Helpdesk s Anthony 6 June 2006 Supervisors: Dr. Yuval Marom Dr. David Albrecht.
Duyu Tang, Furu Wei, Nan Yang, Ming Zhou, Ting Liu, Bing Qin
Learning Table Extraction from Examples Ashwin Tengli, Yiming Yang and Nian Li Ma School of Computer Science Carnegie Mellon University Coling 04.
Extracting Opinions, Opinion Holders, and Topics Expressed in Online News Media Text Soo-Min Kim and Eduard Hovy USC Information Sciences Institute 4676.
Mining and Summarizing Customer Reviews
Opinion mining in social networks Student: Aleksandar Ponjavić 3244/2014 Mentor: Profesor dr Veljko Milutinović.
AQUAINT Kickoff Meeting – December 2001 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification on Reviews Peter D. Turney Institute for Information Technology National.
Automatic Extraction of Opinion Propositions and their Holders Steven Bethard, Hong Yu, Ashley Thornton, Vasileios Hatzivassiloglou and Dan Jurafsky Department.
Carmen Banea, Rada Mihalcea University of North Texas A Bootstrapping Method for Building Subjectivity Lexicons for Languages.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
Bayesian Networks. Male brain wiring Female brain wiring.
1 A study on automatically extracted keywords in text categorization Authors:Anette Hulth and Be´ata B. Megyesi From:ACL 2006 Reporter: 陳永祥 Date:2007/10/16.
A Compositional Context Sensitive Multi-document Summarizer: Exploring the Factors That Influence Summarization Ani Nenkova, Stanford University Lucy Vanderwende,
Distributional Part-of-Speech Tagging Hinrich Schütze CSLI, Ventura Hall Stanford, CA , USA NLP Applications.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
Eric H. Huang, Richard Socher, Christopher D. Manning, Andrew Y. Ng Computer Science Department, Stanford University, Stanford, CA 94305, USA ImprovingWord.
This work is supported by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior National Business Center contract number.
Exploiting Subjectivity Classification to Improve Information Extraction Ellen Riloff University of Utah Janyce Wiebe University of Pittsburgh William.
A Language Independent Method for Question Classification COLING 2004.
Partially Supervised Classification of Text Documents by Bing Liu, Philip Yu, and Xiaoli Li Presented by: Rick Knowles 7 April 2005.
Learning from Multi-topic Web Documents for Contextual Advertisement KDD 2008.
DISCRIMINATIVE TRAINING OF LANGUAGE MODELS FOR SPEECH RECOGNITION Hong-Kwang Jeff Kuo, Eric Fosler-Lussier, Hui Jiang, Chin-Hui Lee ICASSP 2002 Min-Hsuan.
Recognizing Names in Biomedical Texts: a Machine Learning Approach GuoDong Zhou 1,*, Jie Zhang 1,2, Jian Su 1, Dan Shen 1,2 and ChewLim Tan 2 1 Institute.
A Bootstrapping Method for Building Subjectivity Lexicons for Languages with Scarce Resources Author: Carmen Banea, Rada Mihalcea, Janyce Wiebe Source:
Opinion Mining of Customer Feedback Data on the Web Presented By Dongjoo Lee, Intelligent Databases Systems Lab. 1 Dongjoo Lee School of Computer Science.
Noun-Phrase Analysis in Unrestricted Text for Information Retrieval David A. Evans, Chengxiang Zhai Laboratory for Computational Linguistics, CMU 34 th.
1 Multi-Perspective Question Answering Using the OpQA Corpus (HLT/EMNLP 2005) Veselin Stoyanov Claire Cardie Janyce Wiebe Cornell University University.
CISC Machine Learning for Solving Systems Problems Presented by: Ashwani Rao Dept of Computer & Information Sciences University of Delaware Learning.
Automatic Identification of Pro and Con Reasons in Online Reviews Soo-Min Kim and Eduard Hovy USC Information Sciences Institute Proceedings of the COLING/ACL.
CSKGOI'08 Commonsense Knowledge and Goal Oriented Interfaces.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
Recognizing Stances in Online Debates Unsupervised opinion analysis method for debate-side classification. Mine the web to learn associations that are.
Multi-level Bootstrapping for Extracting Parallel Sentence from a Quasi-Comparable Corpus Pascale Fung and Percy Cheung Human Language Technology Center,
Probabilistic Text Structuring: Experiments with Sentence Ordering Mirella Lapata Department of Computer Science University of Sheffield, UK (ACL 2003)
Learning Subjective Nouns using Extraction Pattern Bootstrapping Ellen Riloff School of Computing University of Utah Janyce Wiebe, Theresa Wilson Computing.
Improved Video Categorization from Text Metadata and User Comments ACM SIGIR 2011:Research and development in Information Retrieval - Katja Filippova -
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
Finding document topics for improving topic segmentation Source: ACL2007 Authors: Olivier Ferret (18 route du Panorama, BP6) Reporter:Yong-Xiang Chen.
Discovering Relations among Named Entities from Large Corpora Takaaki Hasegawa *, Satoshi Sekine 1, Ralph Grishman 1 ACL 2004 * Cyberspace Laboratories.
SENTIWORDNET: A Publicly Available Lexical Resource for Opinion Mining
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
Extracting and Ranking Product Features in Opinion Documents Lei Zhang #, Bing Liu #, Suk Hwan Lim *, Eamonn O’Brien-Strain * # University of Illinois.
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
Virtual Examples for Text Classification with Support Vector Machines Manabu Sassano Proceedings of the 2003 Conference on Emprical Methods in Natural.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Annotating and measuring Temporal relations in texts Philippe Muller and Xavier Tannier IRIT,Université Paul Sabatier COLING 2004.
Instance Discovery and Schema Matching With Applications to Biological Deep Web Data Integration Tantan Liu, Fan Wang, Gagan Agrawal {liut, wangfa,
BAYESIAN LEARNING. 2 Bayesian Classifiers Bayesian classifiers are statistical classifiers, and are based on Bayes theorem They can calculate the probability.
Twitter as a Corpus for Sentiment Analysis and Opinion Mining
Multi-Class Sentiment Analysis with Clustering and Score Representation Yan Zhu.
Data Mining and Text Mining. The Standard Data Mining process.
Opinion spam and Analysis 소프트웨어공학 연구실 G 최효린 1 / 35.
An Effective Statistical Approach to Blog Post Opinion Retrieval Ben He, Craig Macdonald, Jiyin He, Iadh Ounis (CIKM 2008)
Unsupervised Learning Part 2. Topics How to determine the K in K-means? Hierarchical clustering Soft clustering with Gaussian mixture models Expectation-Maximization.
Language Identification and Part-of-Speech Tagging
Korean version of GloVe Applying GloVe & word2vec model to Korean corpus speaker : 양희정 date :
Presentation transcript:

2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying the Polarity of Opinion Sentences Lee jae jeong Software Engineering Laboratory, School of Computer Science, University of Seoul

S E Software Engineering Laboratory, School of Computer Science 2 1 Introduction  Newswire articles include those that mainly present opinions or ideas.  Text materials from many other sources contain mixed facts and opinions.  Our motivation for building the opinion detection and classification system described in this paper is the need for organizing information in the context of question answering for complex questions.  such as “What are the reasons for the US-Iraq war?” require long answers that must be constructed from multiple sources.  In such a context, it is imperative that the question answering system can discriminate between opinions and facts, and either use the appropriate type depending on the question or combine them in a meaningful presentation.  Fully analyzing and classifying opinions  These include not only recognizing that the text is subjective, but also determining who the holder of the opinion is, what the opinion is about.  In this paper, we are presenting three of the components of our opinion detection and organization subsystem, which have already been integrated into our larger question-answering system.

S E Software Engineering Laboratory, School of Computer Science 3 2 Related Work  Wiebe et al. (1999) described a sentence-level Naive Bayes classifier using as features the presence or absence of particular syntactic classes (pronouns, adjectives, cardinal numbers, modal verbs, adverbs), punctuation, and sentence position.  Wiebe (2000) introduced lexical features in addition to the presence/absence of syntactic categories.  More recently, Wiebe et al. (2002) report on document-level subjectivity classification, using a k-nearest neighbor algorithm based on the total count of subjective words and phrases within each document.  Our approach to document and sentence classification of opinions builds upon the earlier work by using extended lexical models with additional features.

S E Software Engineering Laboratory, School of Computer Science 4 2 Related Work  Unlike the work cited above, we do not rely on human annotations for training but only on weak metadata provided at the document level.  Our sentence-level classifiers introduce additional criteria for detecting subjective material (opinions), including methods based on sentence similarity within a topic and an approach that relies on multiple classifiers.  For determining whether an opinion sentence is positive or negative, we have used seed words similar to those produced by (Hatzivassiloglou and McKeown, 1997).  Hatzivassiloglou and McKeown (1997) described an unsupervised learning method for obtaining positively and negatively oriented adjectives with accuracy over 90%,

S E Software Engineering Laboratory, School of Computer Science 5 3 Document Classification  To separate documents that contain primarily opinions from documents that report mainly facts, we applied Naive Bayes, a commonly used supervised machine-learning algorithm.  This approach presupposes the availability of at least a collection of articles with pre-assigned opinion and fact labels at the document level;  fortunately, Wall Street Journal articles contain such metadata by identifying the type of each article as Editorial, Letter to editor, Business and News.

S E Software Engineering Laboratory, School of Computer Science 6 4 Finding Opinion Sentences  We developed three different approaches to classify opinions from facts at the sentence level.  4.1 Similarity Approach  Our first approach to classifying sentences as opinions or facts explores the hypothesis that, within a given topic, opinion sentences will be more similar to other opinion sentences than to factual sentences.  We used SIMFINDER (Hatzivassiloglou et al., 2001), a state-of-the-art system for measuring sentence similarity.  To measure the overall similarity of a sentence to the opinion or fact documents, we first select the documents that are on the same topic as the sentence in question.  We then average its SIMFINDER-provided similarities with each sentence in those documents.  Then we assign the sentence to the category for which the average is higher (we call this approach the “score” variant).

S E Software Engineering Laboratory, School of Computer Science 7 4 Finding Opinion Sentences  4.2 Naive Bayes Classifier  Our second method trains a Naive Bayes classifier (see Section 3), using the sentences in opinion and fact documents as the examples of the two categories.  The features include words, bigrams, and trigrams, as well as the parts of speech in each sentence.  In addition, the presence of semantically oriented (positive and negative) words in a sentence is an indicator that the sentence is subjective Hatzivassiloglou and Wiebe, 2000).  Therefore, we include in our features the counts of positive and negative words in the sentence (which are obtained with the method of Section 5.1), as well as counts of the polarities of sequences of semantically oriented words (e.g., “++” for two consecutive positively oriented words).

S E Software Engineering Laboratory, School of Computer Science 8 4 Finding Opinion Sentences  4.3 Multiple Naive Bayes Classifiers  we apply an algorithm using multiple classifiers, each relying on a different subset of our features.  The goal is to reduce the training set to the sentences that are most likely to be correctly labeled, thus boosting classification accuracy.  Given separate sets of features we train separate Naive Bayes classifier corresponding to each feature set.  Assuming as ground truth the information provided by the document labels and that all sentences inherit the status of their document as opinions or facts, we first train C 1 on the entire training set, then use the resulting classifier to predict labels for the training set.  The sentences that receive a label different from the assumed truth are then removed, and we train C 2 on the remaining sentences.  This process is repeated iteratively until no more sentences can be removed.

S E Software Engineering Laboratory, School of Computer Science 9 5 Identifying the Polarity of Opinion Sentences  Having distinguished whether a sentence is a fact or opinion, we separate positive, negative, and neutral opinions into three classes.  We base this decision on the number and strength of semantically oriented words (either positive or negative) in the sentence.  5.1 Semantically Oriented Words  To determine which words are semantically oriented, in what direction, and the strength of their orientation, we measured their co-occurrence with words from a known seed set of semantically oriented words.  The approach is based on the hypothesis that positive words co-occur more than expected by chance, and so do negative words.  As seed words, we used subsets of the 1,336 adjectives that were manually classified as positive (657) or negative (679) by Hatzivassiloglou and McKeown (1997).

S E Software Engineering Laboratory, School of Computer Science 10 5 Identifying the Polarity of Opinion Sentences  5.1 Semantically Oriented Words  For a given seed set size, we denote the set of positive seeds as ADJ p and the set of negative seeds as ADJ n.  We then calculate a modified log-likelihood ratio for a word w i with part of speech POS j (j can be adjective, adverb, noun or verb) as the ratio of its collocation frequency with ADJ p and ADJ n within a sentence,  where represents the collocation frequency of all words wall of part of speech POS j with ADJ p.

S E Software Engineering Laboratory, School of Computer Science 11 6 Data  We used the TREC 8, 9, and 11 collections, which consist of more than 1.7 million newswire articles.  The aggregate collection covers six different newswire sources including 173,252 Wall Street Journal (WSJ) articles from 1987 to  We randomly selected 2,000 articles from each category so that our data set was approximate evenly divided between fact and opinion articles.  Those articles were used for both document and sentence level opinion/fact classification.

S E Software Engineering Laboratory, School of Computer Science 12 7 Evaluation Metrics and Gold Standard  We evaluated the quality of semantically oriented words by mapping the extracted words and labels to an external gold standard.  We mapped article types News and Business to facts, and article types Editorial and Letter to the Editor to opinions.  we created two gold standards for sentence classification.  The first (Standard A) includes the 300 sentences with one judgment and a single judgment for the remaining 100 sentences.  The second standard (Standard B) contains the subset of the 100 sentences for which we obtained identical labels.

S E Software Engineering Laboratory, School of Computer Science 13 8 Results and Discussion  Document Classification We trained our Bayes classifier for documents on 4,000 articles from the WSJ portion of our combined TREC collection, and evaluated on 4,000 other articles also from the WSJ part.  Table 2 lists the F-measure scores (the harmonic mean of precision and recall) of our Bayesian classifier for document-level opinion/fact classification.  The results show the classifier achieved 97% F-measure, which is comparable or higher than the 93% accuracy reported by (Wiebe et al., 2002), who evaluated their work based on a similar set of WSJ articles.

S E Software Engineering Laboratory, School of Computer Science 14 8 Results and Discussion  Sentence Classification Table 3 shows the recall and precision of the similarity-based approach, while Table 4 lists the recall and precision of naïve Bayes (single and multiple classifiers) for sentence level opinion/fact classification.  In both cases, the results are better when we evaluate against Standard B, containing the sentences for which two humans assign the same label.

S E Software Engineering Laboratory, School of Computer Science 15 8 Results and Discussion  Polarity Classification Using the method of Section 5.1, we automatically identified a total of 39,652 (65,773), 3,128 (4,426), 144,238 (195,984), and 22,279 (30,609) positive (negative) adjectives, adverbs, nouns, and verbs.  Extracted positive words include inspirational, truly, luck, and achieve. Negative ones include depraved, disastrously, problem, and depress.  Figure 1 plots the recall and precision of extracted adjectives by using randomly selected seed sets of 1, 20, and 100 pairs of positive and negative adjectives from the list of (Hatzivassiloglou and McKeown, 1997).

S E Software Engineering Laboratory, School of Computer Science 16 9 Conclusions  We presented several models for distinguishing between opinions and facts, and between positive and negative opinions.  At the document level, a fairly straightforward Bayesian classifier using lexical information can distinguish between mostly factual and mostly opinion documents with very high precision and recall (F-measure of 97%).  The task is much harder at the sentence level.  For that case, we described three novel techniques for opinion/fact classification achieving up to 91% precision and recall on the detection of opinion sentences.