Recognizing Contextual Polarity in Phrase-Level Sentiment Analysis Theresa Wilson Janyce Wiebe Paul Hoffmann University of Pittsburgh.

Slides:



Advertisements
Similar presentations
GermanPolarityClues A Lexical Resource for German Sentiment Analysis
Advertisements

Farag Saad i-KNOW 2014 Graz- Austria,
Subjectivity and Sentiment Analysis of Arabic Tweets with Limited Resources Supervisor Dr. Verena Rieser Presented By ESHRAG REFAEE OSACT 27 May 2014.
TEMPLATE DESIGN © Identifying Noun Product Features that Imply Opinions Lei Zhang Bing Liu Department of Computer Science,
LINGUISTICA GENERALE E COMPUTAZIONALE SENTIMENT ANALYSIS.
Playing the Telephone Game: Determining the Hierarchical Structure of Perspective and Speech Expressions Eric Breck and Claire Cardie Department of Computer.
Jan Wiebe University of Pittsburgh Claire Cardie Cornell University Ellen Riloff University of Utah Opinions in Question Answering.
Text Categorization Moshe Koppel Lecture 8: Bottom-Up Sentiment Analysis Some slides adapted from Theresa Wilson and others.
Extract from various presentations: Bing Liu, Aditya Joshi, Aster Data … Sentiment Analysis January 2012.
Sentiment Analysis An Overview of Concepts and Selected Techniques.
A Brief Overview. Contents Introduction to NLP Sentiment Analysis Subjectivity versus Objectivity Determining Polarity Statistical & Linguistic Approaches.
Approaches to Sentiment Analysis MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way Based in part on notes from Aditya Joshi.
Annotating Topics of Opinions Veselin Stoyanov Claire Cardie.
Context-Enhanced Citation Sentiment Analysis Awais Athar & Simone Teufel.
Joint Sentiment/Topic Model for Sentiment Analysis Chenghua Lin & Yulan He CIKM09.
Information Extraction Lecture 11 – Sentiment Analysis CIS, LMU München Winter Semester Dr. Alexander Fraser, CIS.
A Sentimental Education: Sentiment Analysis Using Subjectivity Summarization Based on Minimum Cuts 04 10, 2014 Hyun Geun Soo Bo Pang and Lillian Lee (2004)
Multi-Perspective Question Answering Using the OpQA Corpus Veselin Stoyanov Claire Cardie Janyce Wiebe Cornell University University of Pittsburgh.
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
Annotating Expressions of Opinions and Emotions in Language Wiebe, Wilson, Cardie.
Manual and Automatic Subjectivity and Sentiment Analysis Jan Wiebe Josef Ruppenhofer Swapna Somasundaran University of Pittsburgh.
Predicting the Semantic Orientation of Adjective Vasileios Hatzivassiloglou and Kathleen R. McKeown Presented By Yash Satsangi.
1 Attributions and Private States Jan Wiebe (U. Pittsburgh) Theresa Wilson (U. Pittsburgh) Claire Cardie (Cornell U.)
Learning Subjective Nouns using Extraction Pattern Bootstrapping Ellen Riloff, Janyce Wiebe, Theresa Wilson Presenter: Gabriel Nicolae.
Boosting Applied to Tagging and PP Attachment By Aviad Barzilai.
Learning Subjective Adjectives from Corpora Janyce M. Wiebe Presenter: Gabriel Nicolae.
Extracting Interest Tags from Twitter User Biographies Ying Ding, Jing Jiang School of Information Systems Singapore Management University AIRS 2014, Kuching,
Analyzing Sentiment in a Large Set of Web Data while Accounting for Negation AWIC 2011 Bas Heerschop Erasmus School of Economics Erasmus University Rotterdam.
Extracting Opinions, Opinion Holders, and Topics Expressed in Online News Media Text Soo-Min Kim and Eduard Hovy USC Information Sciences Institute 4676.
PNC 2011: Pacific Neighborhood Consortium S-Sense: An Opinion Mining Tool for Market Intelligence Choochart Haruechaiyasak and Alisa Kongthon Speech and.
Opinion mining in social networks Student: Aleksandar Ponjavić 3244/2014 Mentor: Profesor dr Veljko Milutinović.
Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification on Reviews Peter D. Turney Institute for Information Technology National.
Automatic Extraction of Opinion Propositions and their Holders Steven Bethard, Hong Yu, Ashley Thornton, Vasileios Hatzivassiloglou and Dan Jurafsky Department.
Carmen Banea, Rada Mihalcea University of North Texas A Bootstrapping Method for Building Subjectivity Lexicons for Languages.
A Holistic Lexicon-Based Approach to Opinion Mining Xiaowen Ding, Bing Liu and Philip Yu Department of Computer Science University of Illinois at Chicago.
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
Word Sense and Subjectivity Jan Wiebe Rada Mihalcea University of Pittsburgh University of North Texas.
A Weakly-Supervised Approach to Argumentative Zoning of Scientific Documents Yufan Guo Anna Korhonen Thierry Poibeau 1 Review By: Pranjal Singh Paper.
Exploiting Subjectivity Classification to Improve Information Extraction Ellen Riloff University of Utah Janyce Wiebe University of Pittsburgh William.
Learning from Multi-topic Web Documents for Contextual Advertisement KDD 2008.
A Bootstrapping Method for Building Subjectivity Lexicons for Languages with Scarce Resources Author: Carmen Banea, Rada Mihalcea, Janyce Wiebe Source:
Subjectivity and Sentiment Analysis Jan Wiebe Department of Computer Science CERATOPS: Center for the Extraction and Summarization of Events and Opinions.
Subjectivity and Sentiment Analysis Jan Wiebe Department of Computer Science CERATOPS: Center for the Extraction and Summarization of Events and Opinions.
1 Multi-Perspective Question Answering Using the OpQA Corpus (HLT/EMNLP 2005) Veselin Stoyanov Claire Cardie Janyce Wiebe Cornell University University.
Summarization Focusing on Polarity or Opinion Fragments in Blogs Yohei Seki Toyohashi University of Technology Visiting Scholar at Columbia University.
Predicting Student Emotions in Computer-Human Tutoring Dialogues Diane J. Litman&Kate Forbes-Riley University of Pittsburgh Department of Computer Science.
Automatic Identification of Pro and Con Reasons in Online Reviews Soo-Min Kim and Eduard Hovy USC Information Sciences Institute Proceedings of the COLING/ACL.
Recognizing Stances in Ideological Online Debates.
Emotions from text: machine learning for text-based emotion prediction Cecilia Alm, Dan Roth, Richard Sproat UIUC, Illinois HLT/EMPNLP 2005.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
CSC 594 Topics in AI – Text Mining and Analytics
Evaluating an Opinion Annotation Scheme Using a New Multi- perspective Question and Answer Corpus (AAAI 2004 Spring) Veselin Stoyanov Claire Cardie Diane.
Learning Subjective Nouns using Extraction Pattern Bootstrapping Ellen Riloff School of Computing University of Utah Janyce Wiebe, Theresa Wilson Computing.
11 Project, Part 3. Outline Basics of supervised learning using Naïve Bayes (using a simpler example) Features for the project 2.
7/2003EMNLP031 Learning Extraction Patterns for Subjective Expressions Ellen Riloff Janyce Wiebe University of Utah University of Pittsburgh.
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
Sentiment and Opinion Sep13, 2012 Analysis of Social Media Seminar William Cohen.
Extracting Opinion Topics for Chinese Opinions using Dependence Grammar Guang Qiu, Kangmiao Liu, Jiajun Bu*, Chun Chen, Zhiming Kang Reporter: Chia-Ying.
Learning Event Durations from Event Descriptions Feng Pan, Rutu Mulkar, Jerry R. Hobbs University of Southern California ACL ’ 06.
Word Sense and Subjectivity (Coling/ACL 2006) Janyce Wiebe Rada Mihalcea University of Pittsburgh University of North Texas Acknowledgements: This slide.
Subjectivity and Sentiment Analysis Jan Wiebe Department of Computer Science Intelligent Systems Program University of Pittsburgh.
Finding strong and weak opinion clauses Theresa Wilson, Janyce Wiebe, Rebecca Hwa University of Pittsburgh Just how mad are you? AAAI-2004.
Twitter as a Corpus for Sentiment Analysis and Opinion Mining
Multi-Class Sentiment Analysis with Clustering and Score Representation Yan Zhu.
Identifying Expressions of Opinion in Context Eric Breck and Yejin Choi and Claire Cardie IJCAI 2007.
Sentiment analysis algorithms and applications: A survey
Aspect-based sentiment analysis
Quanzeng You, Jiebo Luo, Hailin Jin and Jianchao Yang
Text Mining & Natural Language Processing
Presentation transcript:

Recognizing Contextual Polarity in Phrase-Level Sentiment Analysis Theresa Wilson Janyce Wiebe Paul Hoffmann University of Pittsburgh

HLT-EMNLP Introduction Sentiment analysis task of identifying positive and negative opinions, emotions, and evaluations  How detailed? Depends on the application. Flame detection, review classification  document-level analysis Question answering, review mining  sentence or phrase-level analysis

HLT-EMNLP Question Answering Example African observers generally approved of his victory while Western Governments denounced it. Q: What is the international reaction to the reelection of Robert Mugabe as President of Zimbabwe?

HLT-EMNLP Most approaches use a lexicon of positive and negative words Prior polarity: out of context, positive or negative beautiful  positive horrid  negative A word may appear in a phrase that expresses a different polarity in context Contextual polarity “Cheers to Timothy Whitfield for the wonderfully horrid visuals.” Prior Polarity versus Contextual Polarity

HLT-EMNLP Example Philip Clap, President of the National Environment Trust, sums up well the general thrust of the reaction of environmental movements: there is no reason at all to believe that the polluters are suddenly going to become reasonable.

HLT-EMNLP Example Philip Clap, President of the National Environment Trust, sums up well the general thrust of the reaction of environmental movements: there is no reason at all to believe that the polluters are suddenly going to become reasonable.

HLT-EMNLP Philip Clap, President of the National Environment Trust, sums up well the general thrust of the reaction of environmental movements: there is no reason at all to believe that the polluters are suddenly going to become reasonable. Example prior polarity Contextual polarity

HLT-EMNLP Goal of Our Research Automatically distinguish prior and contextual polarity

HLT-EMNLP Approach Use machine learning and variety of features Achieve significant results for a large subset of sentiment expressions Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances

HLT-EMNLP Outline Introduction Manual Annotations Corpus Prior-Polarity Subjectivity Lexicon Experiments Previous Work Conclusions

HLT-EMNLP Manual Annotations Need: sentiment expressions with contextual polarity  positive and negative expressions of emotions, evaluations, stances Had: subjective expression annotations in MPQA Opinion Corpus  words/phrases expressing emotions, evaluations, stances, speculations, etc. sentiment expressions  subjective expressions Decision: annotate subjective expressions in MPQA Corpus with their contextual polarity

HLT-EMNLP Annotation Scheme Mark polarity of subjective expressions as positive, negative, both, or neutral African observers generally approved of his victory while Western governments denounced it. Besides, politicians refer to good and evil … Jerome says the hospital feels no different than a hospital in the states. positive negative both neutral

HLT-EMNLP Annotation Scheme Judge the contextual polarity of sentiment ultimately being conveyed They have not succeeded, and will never succeed, in breaking the will of this valiant people. positive

HLT-EMNLP Agreement Study 10 documents with 447 subjective expressions Kappa: 0.72 (82%) Remove uncertain cases  at least one annotator marked uncertain (18%) Kappa: 0.84 (90%) (But all data included in experiments)

HLT-EMNLP Outline Introduction Manual Annotations Corpus Prior-Polarity Subjectivity Lexicon Experiments Previous Work Conclusions Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances

HLT-EMNLP Corpus 425 documents from MPQA Opinion Corpus 15,991 subjective expressions in 8,984 sentences Divided into two sets Development set 66 docs / 2,808 subjective expressions Experiment set 359 docs / 13,183 subjective expressions Divided into 10 folds for cross-validation

HLT-EMNLP Outline Introduction Manual Annotations Corpus Prior-Polarity Subjectivity Lexicon Experiments Previous Work Conclusions Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances

HLT-EMNLP Prior-Polarity Subjectivity Lexicon Over 8,000 words from a variety of sources Both manually and automatically identified Positive/negative words from General Inquirer and Hatzivassiloglou and McKeown (1997) All words in lexicon tagged with: Prior polarity: positive, negative, both, neutral Reliability: strongly subjective (strongsubj), weakly subjective (weaksubj)

HLT-EMNLP Outline Introduction Manual Annotations Corpus Prior-Polarity Subjectivity Lexicon Experiments Previous Work Conclusions

HLT-EMNLP Experiments  Give each instance its own label Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances Both Steps: BoosTexter AdaBoost.HM 5000 rounds boosting 10-fold cross validation

HLT-EMNLP Definition of Gold Standard Given an instance inst from the lexicon: if inst not in a subjective expression: goldclass( inst ) = neutral else if inst in at least one positive and one negative subjective expression: goldclass( inst ) = both else if inst in a mixture of negative and neutral: goldclass( inst ) = negative else if inst in a mixture of positive and neutral: goldclass( inst ) = positive else: goldclass( inst ) = contextual polarity of subjective expression

HLT-EMNLP Features Many inspired by Polanya & Zaenen (2004): Contextual Valence Shifters Example:little threat little truth Others capture dependency relationships between words Example: wonderfully horrid pos mod

HLT-EMNLP Word features 2. Modification features 3. Structure features 4. Sentence features 5. Document feature Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances

HLT-EMNLP Word features 2. Modification features 3. Structure features 4. Sentence features 5. Document feature Word token terrifies Word part-of-speech VB Context that terrifies me Prior Polarity negative Reliability strongsubj Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances

HLT-EMNLP Word features 2. Modification features 3. Structure features 4. Sentence features 5. Document feature Binary features: Preceded by adjective adverb (other than not) intensifier Self intensifier Modifies strongsubj clue weaksubj clue Modified by strongsubj clue weaksubj clue Dependency Parse Tree Thehumanrights report poses asubstantial challenge … det adj mod adj det subjobj p Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances

HLT-EMNLP Word features 2. Modification features 3. Structure features 4. Sentence features 5. Document feature Binary features: In subject The human rights report poses In copular I am confident In passive voice must be regarded Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances Thehumanrights report poses asubstantial challenge … det adj mod adj det subjobj p

HLT-EMNLP Word features 2. Modification features 3. Structure features 4. Sentence features 5. Document feature Count of strongsubj clues in previous, current, next sentence Count of weaksubj clues in previous, current, next sentence Counts of various parts of speech Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances

HLT-EMNLP Word features 2. Modification features 3. Structure features 4. Sentence features 5. Document feature Document topic (15) economics health Kyoto protocol presidential election in Zimbabwe … Example: The disease can be contracted if a person is bitten by a certain tick or if a person comes into contact with the blood of a congo fever sufferer. Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances

HLT-EMNLP Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances Results 1a

HLT-EMNLP Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances Results 1b

HLT-EMNLP Step 2: Polarity Classification Classes positive, negative, both, neutral Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances 19,5065,671

HLT-EMNLP Word token Word prior polarity Negated Negated subject Modifies polarity Modified by polarity Conjunction polarity General polarity shifter Negative polarity shifter Positive polarity shifter Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances

HLT-EMNLP Word token Word prior polarity Negated Negated subject Modifies polarity Modified by polarity Conjunction polarity General polarity shifter Negative polarity shifter Positive polarity shifter Word token terrifies Word prior polarity negative Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances

HLT-EMNLP Word token Word prior polarity Negated Negated subject Modifies polarity Modified by polarity Conjunction polarity General polarity shifter Negative polarity shifter Positive polarity shifter Binary features: Negated For example: not good does not look very good  not only good but amazing Negated subject No politically prudent Israeli could support either of them. Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances

HLT-EMNLP Word token Word prior polarity Negated Negated subject Modifies polarity Modified by polarity Conjunction polarity General polarity shifter Negative polarity shifter Positive polarity shifter Modifies polarity 5 values: positive, negative, neutral, both, not mod substantial: negative Modified by polarity 5 values: positive, negative, neutral, both, not mod challenge: positive substantial (pos) challenge (neg) Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances

HLT-EMNLP Word token Word prior polarity Negated Negated subject Modifies polarity Modified by polarity Conjunction polarity General polarity shifter Negative polarity shifter Positive polarity shifter Conjunction polarity 5 values: positive, negative, neutral, both, not mod good: negative good (pos) and evil (neg) Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances

HLT-EMNLP Word token Word prior polarity Negated Negated subject Modifies polarity Modified by polarity Conjunction polarity General polarity shifter Negative polarity shifter Positive polarity shifter General polarity shifter pose little threat contains little truth Negative polarity shifter lack of understanding Positive polarity shifter abate the damage Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances

HLT-EMNLP Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances Results 2a

HLT-EMNLP Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances Results 2b

HLT-EMNLP Ablation experiments removing features: 1. Negated, negated subject 2. Modifies polarity, modified by polarity 3. Conjunction polarity 4. General, negative, positive polarity shifters Corpus Lexicon Neutral or Polar? Step 1 Contextual Polarity? Step 2 All Instances Polar Instances

HLT-EMNLP Outline Introduction Manual Annotations Corpus Prior-Polarity Subjectivity Lexicon Experiments Previous Work Conclusions

HLT-EMNLP Previous Work Learn prior polarity of words and phrases e.g., Hatzivassiloglou & McKeown (1997), Turney (2002) Sentence-level sentiment analysis e.g., Yu & Hatzivassiloglou (2003), Kim & Hovy (2004) Phrase-level contextual polarity classification e.g., Yi et al. (2003)

HLT-EMNLP At HLT/EMNLP 2005 Popescu & Etizioni: Extracting Product Features and Opinions from Reviews Choi, Cardie, Riloff & Patwardhan: Identifying Sources of Opinions with Conditional Random Fields and Extraction Patterns Alm, Roth & Sproat: Emotions from Text: Machine Learning for Text-based Emotion Prediction

HLT-EMNLP Outline Introduction Manual Annotations Corpus Prior-Polarity Subjectivity Lexicon Experiments Previous Work Conclusions

HLT-EMNLP Conclusions Presented a two-step approach to phrase- level sentiment analysis 1. Determine if an expression is neutral or polar 2. Determines contextual polarity of the ones that are polar Automatically identify the contextual polarity of a large subset of sentiment expression

HLT-EMNLP Thank you

HLT-EMNLP Acknowledgments This work was supported by Advanced Research and Development Activity (ARDA) National Science Foundation