Instructor: Smaranda Muresan Columbia University Sentiment Lexicons.

Slides:



Advertisements
Similar presentations
A Comparison of Implicit and Explicit Links for Web Page Classification Dou Shen 1 Jian-Tao Sun 2 Qiang Yang 1 Zheng Chen 2 1 Department of Computer Science.
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Farag Saad i-KNOW 2014 Graz- Austria,
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
Distant Supervision for Emotion Classification in Twitter posts 1/17.
Sentiment Analysis Learning Sentiment Lexicons. Dan Jurafsky Semi-supervised learning of lexicons Use a small amount of information A few labeled examples.
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Computational Models of Discourse Analysis Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
TEMPLATE DESIGN © Identifying Noun Product Features that Imply Opinions Lei Zhang Bing Liu Department of Computer Science,
LINGUISTICA GENERALE E COMPUTAZIONALE SENTIMENT ANALYSIS.
Text Categorization Moshe Koppel Lecture 8: Bottom-Up Sentiment Analysis Some slides adapted from Theresa Wilson and others.
Extract from various presentations: Bing Liu, Aditya Joshi, Aster Data … Sentiment Analysis January 2012.
Sentiment Analysis An Overview of Concepts and Selected Techniques.
CIS630 Spring 2013 Lecture 2 Affect analysis in text and speech.
A Sentimental Education: Sentiment Analysis Using Subjectivity Summarization Based on Minimum Cuts 04 10, 2014 Hyun Geun Soo Bo Pang and Lillian Lee (2004)
CS Word Sense Disambiguation. 2 Overview A problem for semantic attachment approaches: what happens when a given lexeme has multiple ‘meanings’?
Sentiment Lexicon Creation from Lexical Resources BIS 2011 Bas Heerschop Erasmus School of Economics Erasmus University Rotterdam
Predicting the Semantic Orientation of Adjective Vasileios Hatzivassiloglou and Kathleen R. McKeown Presented By Yash Satsangi.
Predicting the Semantic Orientation of Adjectives
Learning Subjective Adjectives from Corpora Janyce M. Wiebe Presenter: Gabriel Nicolae.
Seven Lectures on Statistical Parsing Christopher Manning LSA Linguistic Institute 2007 LSA 354 Lecture 7.
1 Extracting Product Feature Assessments from Reviews Ana-Maria Popescu Oren Etzioni
M. Sulaiman Khan Dept. of Computer Science University of Liverpool 2009 COMP527: Data Mining Classification: Evaluation February 23,
Mining and Summarizing Customer Reviews
Mining the Peanut Gallery: Opinion Extraction and Semantic Classification of Product Reviews K. Dave et al, WWW 2003, citations Presented by Sarah.
(ACM KDD 09’) Prem Melville, Wojciech Gryc, Richard D. Lawrence
Mining and Summarizing Customer Reviews Minqing Hu and Bing Liu University of Illinois SIGKDD 2004.
Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification on Reviews Peter D. Turney Institute for Information Technology National.
Carmen Banea, Rada Mihalcea University of North Texas A Bootstrapping Method for Building Subjectivity Lexicons for Languages.
Opinion Mining Using Econometrics: A Case Study on Reputation Systems Anindya Ghose, Panagiotis G. Ipeirotis, and Arun Sundararajan Leonard N. Stern School.
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
Sentiment Analysis What is Sentiment Analysis?. Positive or negative movie review? unbelievably disappointing Full of zany characters and richly applied.
This work is supported by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior National Business Center contract number.
Sentiment Detection Naveen Sharma( ) PrateekChoudhary( ) Yashpal Meena( ) Under guidance Of Prof. Pushpak Bhattacharya.
Learning from Multi-topic Web Documents for Contextual Advertisement KDD 2008.
CS 4705 Lecture 19 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised.
Opinion Mining of Customer Feedback Data on the Web Presented By Dongjoo Lee, Intelligent Databases Systems Lab. 1 Dongjoo Lee School of Computer Science.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Ensemble Learning Spring 2009 Ben-Gurion University of the Negev.
A Scalable Machine Learning Approach for Semi-Structured Named Entity Recognition Utku Irmak(Yahoo! Labs) Reiner Kraft(Yahoo! Inc.) WWW 2010(Information.
TEXT ANALYTICS - LABS Maha Althobaiti Udo Kruschwitz Massimo Poesio.
Slide 10-1 © 1999 South-Western Publishing McDaniel Gates Contemporary Marketing Research, 4e Using Measurement Scales to Build Marketing Effectiveness.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
Extracting Hidden Components from Text Reviews for Restaurant Evaluation Juanita Ordonez Data Mining Final Project Instructor: Dr Shahriar Hossain Computer.
Recognizing Stances in Online Debates Unsupervised opinion analysis method for debate-side classification. Mine the web to learn associations that are.
Machine Learning Tutorial-2. Recall, Precision, F-measure, Accuracy Ch. 5.
Natural Language Processing TJTSD66: Advanced Topics in Social Media Dr. WANG, CS & IS, JYU
SENTIWORDNET: A Publicly Available Lexical Resource for Opinion Mining
Show Me the Money! Deriving the Pricing Power of Product Features by Mining Consumer Reviews Nikolay Archak, Anindya Ghose, and Panagiotis G. Ipeirotis.
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
Extracting and Ranking Product Features in Opinion Documents Lei Zhang #, Bing Liu #, Suk Hwan Lim *, Eamonn O’Brien-Strain * # University of Illinois.
Sentiment and Opinion Sep13, 2012 Analysis of Social Media Seminar William Cohen.
Notes on HW 1 grading I gave full credit as long as you gave a description, confusion matrix, and working code Many people’s descriptions were quite short.
LECTURE 02: EVALUATING MODELS January 27, 2016 SDS 293 Machine Learning.
Learning Event Durations from Event Descriptions Feng Pan, Rutu Mulkar, Jerry R. Hobbs University of Southern California ACL ’ 06.
Naïve Bayes Classifier April 25 th, Classification Methods (1) Manual classification Used by Yahoo!, Looksmart, about.com, ODP Very accurate when.
Sentiment Analysis Using Common- Sense and Context Information Basant Agarwal 1,2, Namita Mittal 2, Pooja Bansal 2, and Sonal Garg 2 1 Department of Computer.
Semi-Supervised Recognition of Sarcastic Sentences in Twitter and Amazon -Smit Shilu.
Twitter as a Corpus for Sentiment Analysis and Opinion Mining
A Document-Level Sentiment Analysis Approach Using Artificial Neural Network and Sentiment Lexicons Yan Zhu.
Opinion spam and Analysis 소프트웨어공학 연구실 G 최효린 1 / 35.
Sentiment Analysis of Twitter Messages Using Word2Vec
Sentiment Analysis Seminar Social Media Mining University UC3M
Erasmus University Rotterdam
What is Sentiment Analysis?
Aspect-based sentiment analysis
An Overview of Concepts and Selected Techniques
Introduction to Sentiment Analysis
Presentation transcript:

Instructor: Smaranda Muresan Columbia University Sentiment Lexicons

Announcements Class setup on Courseworks too. Class website linked to Courseworks (“Class website” tab) TA’s (Arpit Gupta) office hours – Monday 4:15-5:15pm in TA room in Mudd TA’s –

Class Today Word level sentiment analysis (Sentiment Lexicons) Discussion of the two papers Introduction to Sentiment Analysis beyond words (sentence level, text level) (to facilitate discussion of articles next week)

What is sentiment analysis? Attempts to identify the sentiment/opinion that a person may hold towards an object/person/topic etc It is a finer grain analysis compared to subjectivity analysis Sentiment AnalysisSubjectivity analysis Positive Subjective Negative NeutralObjective This film should be brilliant. It sounds like a great plot, the actors are first grade, and the supporting cast is good as well, and Stallone is attempting to deliver a good performance. However, it can’t hold up.

Why sentiment analysis? Movie: is this review positive or negative? Products: what do people think about the new iPhone? Public sentiment: how is consumer confidence? Is despair increasing? Politics: what do people think about this candidate or issue? Prediction: predict election outcomes or market trends from sentiment 5

Goal of today’s lecture Gain insights into how sentiment is expressed lexically Begin developing resources that are useful in higher level classification (phrase level, sentence level, document level) Explore different philosophies on how to build such large scale sentiment lexicons

What are we classifying gross (gross,adj) (gross,noun) (gross,verb) gross out GROSS!!! The soup was gross – 1 star The horror movie was gross – 5 stars

Words Adjectives – positive: honest important mature large patient Ron Paul is the only honest man in Washington. Kitchell’s writing is unbelievably mature and is only likely to get better. To humour me my patient father agrees yet again to my choice of film

Words Adjectives – negative: harmful hypocritical inefficient insecure It was a macabre and hypocritical circus. Why are they being so inefficient ? Slide from Janyce Wiebe

Other parts of speech Verbs – positive: praise, love – negative: blame, criticize Nouns – positive: pleasure, enjoyment – negative: pain, criticism

Hand Annotated/Compiled Lexicons WordNet-based approaches Distributional Approaches How to build sentiment lexicons

General Inquirer (GI) Harvard General Inquirer Database (Stone, 1966) – Total of 11,788 terms – – – Positive (1915 words) vs Negative (2291 words) (rest of 7582 could be consider Neutral) – Strong vs Weak – Active vs Passive – Overstated versus Understated – Pleasure, Pain, Virtue, Vice – Motivation, Cognitive Orientation, etc

WordNet (Miller, 1995; Fellbaum, 1998) Semantic Lexical resource (multilingual) Synsets (denote different senses of a word)

Micro-WNOp (Cerini et al 1997) 1105 Wordnet Sysnsets related to opinion topic (initial words were selected from the GI)

Micro-WNOp (Carrenini et al 1997) Micro-WNOp statistics reduced to the 702 sysnsets when everyone agreed ISSUES with Hand built Lexicons such as GI, Micro-WNOp???

Hand Annotated/Compiled Lexicons WordNet-based approaches Distributional Approaches How to build sentiment lexicons

Simple sense/sentiment propagation Hypothesis: Sentiment is constant throughout regions of lexically related items. Thus, sentiment properties of hand-built seed-sets will be preserved as we follow WordNet relations out from them. SentiWordNet (Esuli and Sebastiani, 2006) – Approx 1.7 Million words – Using WordNet and Machine Learning (Classifiers). – Each synset is assigned three scores Positive Negative Objective

Values in 3 dimension sum to 1. Ex: P=0.75, N=0, O=0.25

Building SentiWordNet Lp, Ln, Lo are the three seed sets Iteratively expand the seed sets through K steps Train the classifier for the expanded sets

Lp Ln also-see antonymy Expansion of seed sets The sets at the end of kth step are called Tr(k,p) and Tr(k,n) Tr(k,o) is the set that is not present in Tr(k,p) and Tr(k,n)

Committee of classifiers Train a committee of classifiers of different types and different K-values for the given data Observations: – Low values of K give high precision and low recall – Accuracy in determining positivity or negativity, however, remains almost constant

Useful Sentiment Tutorial Has code related to WordNet propagation methods (used in SentiWordNet) Many other pointers! Issues with the WordNet based propagation lexicons?

Other Sentiment Lexicons

MPQA Subjectivity Cues Lexicon Home page: words from 8221 lemmas – 2718 positive – 4912 negative Each word annotated for intensity (strong, weak) GNU GPL 24 Theresa Wilson, Janyce Wiebe, and Paul Hoffmann (2005). Recognizing Contextual Polarity in Phrase-Level Sentiment Analysis. Proc. of HLT-EMNLP Riloff and Wiebe (2003). Learning extraction patterns for subjective expressions. EMNLP-2003.

Bing Liu Opinion Lexicon Bing Liu's Page on Opinion Mining lexicon-English.rar lexicon-English.rar 6786 words – 2006 positive – 4783 negative 25 Minqing Hu and Bing Liu. Mining and Summarizing Customer Reviews. ACM SIGKDD-2004.

Disagreements between polarity lexicons Opinion Lexicon General Inquirer SentiWordNet MPQA33/5402 (0.6%) 49/2867 (2%) 1127/4214 (27%) Opinion Lexicon32/2411 (1%) 1004/3994 (25%) General Inquirer520/2306 (23%) SentiWordNet 26 Christopher Potts, Sentiment Tutorial, 2011Sentiment Tutorial

Hand Annotated/Compiled Lexicons WordNet-based approaches Distributional Approaches – 2 papers for discussion today How to build sentiment lexicons

Presenter: Smaranda Muresan Predicting the semantic orientation of adjectives Hatzivassiloglou & McKeown 1997

Goal Predicting polarity of adjectives from a large corpus Test the hypothesis: the morphosyntactic properties of coordination provide reliable information about adjectival oppositions and lexical polarities

Adjectives conjoined by “and” have same polarity – Fair and legitimate, corrupt and brutal – *fair and brutal, *corrupt and legitimate Adjectives conjoined by “but” do not – fair but brutal

Approach Extract conjunctions of adjectives from a large corpus, along with relevant morphological relations Use a log-linear regression model to predict orientation of two different adjectives Use a clustering algorithm to separate the adjectives into two subsets of different orientation Use average frequencies in each group to assign the label (group with highest frequency is labeled positive)

Seed data Label seed set of 1336 adjectives (all >20 in 21 million word Wall Street Journal corpus) – 657 positive adequate central clever famous intelligent remarkable reputed sensitive slender thriving… – 679 negative contagious drunken ignorant lanky listless primitive strident troublesome unresolved unsuspecting… Further validation: ask 4 human judges to label a subset of 500 adjectives: 96.97% average inter-judge agreement 32

Validating the Hypothesis Run a parser on 21 million words dataset to get 15,048 conjunction tokens involving 9,296 pairs of distinct adjective pairs. Each conjunction was classified into : – 1) conjunction used (and, or, but,…) – 2) type of modification (attributive, predicative) – 3) number modified noun (singular or plural) Considered conjunction where both members were in the seed set (e.g. clever and sensitive) Count percentage of conjunction in each category with adjectives of same or different orientation

Validating Hypothesis For almost all the cases p-values are low. Hence the statistics are significant. ‘and’ usually joins adjectives of same orientation ‘but’ is opposite and joins adjectives of different orientation

Link Prediction classy nice helpful fair brutal irrational corrupt Baseline: always use same orientation – 77.84% the “but” rule morphological rules (adequate-inadequate) Better idea: supervised learning using log-linear regression

Result of Prediction Log Linear Regression models performs slightly better than baseline

Clustering for partitioning the graph into two groups Log Linear model generates a dissimilarity score between two adjective between 0 and 1 37 classy nice helpful fair brutal irrational corrupt

Labeling the clusters Two key insights about pairs of words of opposite orientations: - semantically unmarked member has positive orientation (e.g honest (unmarked) vs dishonest (marked)) - semantically unmarked member is the most frequent 38 classy nice helpful fair brutal irrational corrupt + -

Output polarity lexicon Positive – bold decisive disturbing generous good honest important large mature patient peaceful positive proud sound stimulating straightforward strange talented vigorous witty… Negative – ambiguous cautious cynical evasive harmful hypocritical inefficient insecure irrational irresponsible minor outspoken pleasant reckless risky selfish tedious unsupported vulnerable wasteful… 39

Output polarity lexicon Positive – bold decisive disturbing generous good honest important large mature patient peaceful positive proud sound stimulating straightforward strange talented vigorous witty… Negative – ambiguous cautious cynical evasive harmful hypocritical inefficient insecure irrational irresponsible minor outspoken pleasant reckless risky selfish tedious unsupported vulnerable wasteful… 40

Evaluating Clustering of Adjectives Tried to account for graph connectivity Used the adjectives from seed set (A) and links given by conjunction and morphological rules Separate in training/testing using a parameter α –higher α creates subset of A such that more adjectives are connected to each other.

Clustering Results Highest accuracy obtained when highest number of links were present. Ratio of group frequency correctly identified the positive subgroup

Graph Connectivity and Performance Parameter P measures how well each link is predicted independently – Precision Parameter k – average number of links for each adjective: Goal: even if P is low, given enough data (high k) a high performance for group prediction is achieved

Results

Discussion points What do you see the major contribution of this paper? - Helps to highlight in a quantitative way the relationship between sentiment and particular words and constructions (coordination)- useful linguistic insight - corpus best method (thus avoiding limitation of human built resources such as WordNet) - Can be extended to nouns and verbs. Classic paper, cited 1127 times

Discussion points Does it have all the information for anyone to be able to replicate the results? – How is the dissimilarity value computed? (multiple values are delivered for an adjective pair in different environments) What are the limitations of the approach? – Method is limited by human cleverness in coming up with useful constructions

Velikovich et al

Class Today Word level sentiment analysis (Sentiment Lexicons) Discussion of the two papers Introduction to Sentiment Analysis beyond words (phrase level, text level) (to facilitate discussion of articles next week)

What is sentiment analysis? Attempts to identify the sentiment/opinion/attitude that a person may hold towards an object/person/topic etc

Components 1.Holder (source) of attitude 2.Target (aspect) of attitude 3.Type of attitude From a set of types – Like, love, hate, value, desire, etc. Or (more commonly) simple weighted polarity: – positive, negative, neutral, together with strength 4.Text containing the attitude Sentence or entire document 50 This film should be brilliant. It sounds like a great plot, the actors are first grade, and the supporting cast is good as well, and Stallone is attempting to deliver a good performance. However, it can’t hold up.

Sentiment Analysis Simplest task: – Is the attitude of this text positive or negative? More complex: – Rank the attitude of this text from 1 to 5 Advanced: – Detect the target, source, or complex attitude types

Sentiment Analysis Simplest task: – Is the attitude of this text positive or negative? More complex: – Rank the attitude of this text from 1 to 5 Advanced: – Detect the target, source, or complex attitude types

Sentiment Analysis A Baseline Algorithm

Sentiment Classification in Movie Reviews Polarity detection: – Is an IMDB movie review positive or negative? Data: Polarity Data 2.0: – -review-data -review-data Bo Pang, Lillian Lee, and Shivakumar Vaithyanathan Thumbs up? Sentiment Classification using Machine Learning Techniques. EMNLP-2002, 79—86. Bo Pang and Lillian Lee A Sentimental Education: Sentiment Analysis Using Subjectivity Summarization Based on Minimum Cuts. ACL,

Text Classification: definition The classifier (test phase): – Input: a document d (e.g., a movie review) – Output: a predicted class c from some fixed set of labels c 1,...,c K (e,g,pos, neg) The learner (training phase): – Input: a set of m hand-labeled documents (d 1,c 1 ),....,(d m,c m ) – Output: a learned classifier f:d  c

IMDB data in the Pang and Lee database when _star wars_ came out some twenty years ago, the image of traveling throughout the stars has become a commonplace image. […] when han solo goes light speed, the stars change to bright lines, going towards the viewer in lines that converge at an invisible point. cool. _october sky_ offers a much simpler image–that of a single white dot, traveling horizontally across the night sky. [... ] “ snake eyes ” is the most aggravating kind of movie : the kind that shows so much potential then becomes unbelievably disappointing. it’s not just because this is a brian depalma film, and since he’s a great director and one who’s films are always greeted with at least some fanfare. and it’s not even because this was a film starring nicolas cage and since he gives a brauvara performance, this film is hardly worth his talents. ✓✗

Baseline Algorithm (adapted from Pang and Lee) Tokenization Feature Extraction Classification using different classifiers – Naïve Bayes – MaxEnt – Support Vector Machines (SVM)

Sentiment Tokenization Issues Deal with HTML and XML markup Twitter mark-up (names, hash tags) Capitalization (preserve for words in all caps) Phone numbers, dates Emoticons Useful code: – Christopher Potts sentiment tokenizer Christopher Potts sentiment tokenizer – Brendan O’Connor twitter tokenizer Brendan O’Connor twitter tokenizer 58

Extracting Features for Sentiment Classification How to handle negation –I didn’t like this movie vs –I really like this movie Which words to use? – Only adjectives – All words All words turns out to work better, at least on this data 59

Negation Add NOT_ to every word between negation and following punctuation: didn’t like this movie, but I didn’t NOT_like NOT_this NOT_movie but I Das, Sanjiv and Mike Chen Yahoo! for Amazon: Extracting market sentiment from stock message boards. In Proceedings of the Asia Pacific Finance Association Annual Conference (APFA). Bo Pang, Lillian Lee, and Shivakumar Vaithyanathan Thumbs up? Sentiment Classification using Machine Learning Techniques. EMNLP-2002, 79—86.

Classification methods Naïve Bayes MaxEnt SVM

Evaluating Classification Evaluation must be done on test data that are independent of the training data – usually a disjoint set of instances Classification accuracy: c / n where n is the total number of test instances and c is the number of test instances correctly classified by the system. – Adequate if one class per document Results can vary based on sampling error due to different training and test sets. – Average results over multiple training and test sets (splits of the overall data) for the best results. Slide from Chris Manning

Cross-Validation Break up data into 10 folds – (Equal positive and negative inside each fold?) For each fold – Choose the fold as a temporary test set – Train on 9 folds, compute performance on the test fold Report average performance of the 10 runs

Other issues in Classification MaxEnt and SVM tend to do better than Naïve Bayes 64

Problems: What makes reviews hard to classify? Subtlety: – Perfume review in Perfumes: the Guide: “If you are reading this because it is your darling fragrance, please wear it at home exclusively, and tape the windows shut.” 65

Thwarted Expectations and Ordering Effects “This film should be brilliant. It sounds like a great plot, the actors are first grade, and the supporting cast is good as well, and Stallone is attempting to deliver a good performance. However, it can’t hold up.” Well as usual Keanu Reeves is nothing special, but surprisingly, the very talented Laurence Fishbourne is not so good either, I was surprised. 66

Due Next Class Readings – Chapter 4 from Pang and Lee “Opinion Mining and Sentiment Analysis” book – 2 papers for discussions A short data analysis assignment – Description on Courseworks under Assignments – Goal is to get a better understanding of data and the problems discussed in class – Grade: Excellent/Good/Insufficient – Due before class. No late submissions

Next class Discussion of 2 papers (50 minutes) – 25 minutes per paper – Prepare a 15 min presentations and lead discussion for 10 minutes 5 min break More in depth lecture on sentiment analysis & open questions (can lead to ideas for projects) – 30 minutes Introduction to Emotion/Mood (25 minutes)

Announcements The assignments of paper for discussions will be done by Saturday, Feb 1, 5pm. TA office hours – 4:15-5:15pm Mondays in the TA room in Mudd TA TA if you’d like a tutorial on Text Classification and existing toolkits

Announcements Grading policy slightly updated to include data analysis assignments – 10% data analysis assignments (3 assignments, grading Excellent/Good/Insufficient). No late submissions! See class website or details – 30% discussion of papers – 60% project 10% literature review part 5% class presentation 45% final paper and project