Duyu Tang, Furu Wei, Nan Yang, Ming Zhou, Ting Liu, Bing Qin

Slides:



Advertisements
Similar presentations
A Comparison of Implicit and Explicit Links for Web Page Classification Dou Shen 1 Jian-Tao Sun 2 Qiang Yang 1 Zheng Chen 2 1 Department of Computer Science.
Advertisements

Document Summarization using Conditional Random Fields Dou Shen, Jian-Tao Sun, Hua Li, Qiang Yang, Zheng Chen IJCAI 2007 Hao-Chin Chang Department of Computer.
Entity-Centric Topic-Oriented Opinion Summarization in Twitter Date : 2013/09/03 Author : Xinfan Meng, Furu Wei, Xiaohua, Liu, Ming Zhou, Sujian Li and.
Distant Supervision for Emotion Classification in Twitter posts 1/17.
Supervised Learning Techniques over Twitter Data Kleisarchaki Sofia.
Problem Semi supervised sarcasm identification using SASI
Measuring the Influence of Long Range Dependencies with Neural Network Language Models Le Hai Son, Alexandre Allauzen, Franc¸ois Yvon Univ. Paris-Sud and.
A Survey on Text Categorization with Machine Learning Chikayama lab. Dai Saito.
A Sentimental Education: Sentiment Analysis Using Subjectivity Summarization Based on Minimum Cuts 04 10, 2014 Hyun Geun Soo Bo Pang and Lillian Lee (2004)
1 Noun Homograph Disambiguation Using Local Context in Large Text Corpora Marti A. Hearst Presented by: Heng Ji Mar. 29, 2004.
Distributed Representations of Sentences and Documents
Scalable Text Mining with Sparse Generative Models
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Introduction to Machine Learning Approach Lecture 5.
Longbiao Kang, Baotian Hu, Xiangping Wu, Qingcai Chen, and Yan He Intelligent Computing Research Center, School of Computer Science and Technology, Harbin.
(ACM KDD 09’) Prem Melville, Wojciech Gryc, Richard D. Lawrence
Sentiment-Specific Representation Learning for Document-Level Sentiment Analysis Duyu Tang Research Center for Social Computing and Information Retrieval,
Learning from Imbalanced, Only Positive and Unlabeled Data Yetian Chen
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
Introduction to variable selection I Qi Yu. 2 Problems due to poor variable selection: Input dimension is too large; the curse of dimensionality problem.
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
Analysis of a Neural Language Model Eric Doi CS 152: Neural Networks Harvey Mudd College.
A Scalable Self-organizing Map Algorithm for Textual Classification: A Neural Network Approach to Thesaurus Generation Dmitri G. Roussinov Department of.
Eric H. Huang, Richard Socher, Christopher D. Manning, Andrew Y. Ng Computer Science Department, Stanford University, Stanford, CA 94305, USA ImprovingWord.
1 Co-Training for Cross-Lingual Sentiment Classification Xiaojun Wan ( 萬小軍 ) Associate Professor, Peking University ACL 2009.
Partially Supervised Classification of Text Documents by Bing Liu, Philip Yu, and Xiaoli Li Presented by: Rick Knowles 7 April 2005.
2014 EMNLP Xinxiong Chen, Zhiyuan Liu, Maosong Sun State Key Laboratory of Intelligent Technology and Systems Tsinghua National Laboratory for Information.
TEXT ANALYTICS - LABS Maha Althobaiti Udo Kruschwitz Massimo Poesio.
Presenter: Jinhua Du ( 杜金华 ) Xi’an University of Technology 西安理工大学 NLP&CC, Chongqing, Nov , 2013 Discriminative Latent Variable Based Classifier.
CHAPTER 8 DISCRIMINATIVE CLASSIFIERS HIDDEN MARKOV MODELS.
Neural Text Categorizer for Exclusive Text Categorization Journal of Information Processing Systems, Vol.4, No.2, June 2008 Taeho Jo* 報告者 : 林昱志.
CSC 594 Topics in AI – Text Mining and Analytics
Neural Net Language Models
Multiple Instance Learning for Sparse Positive Bags Razvan C. Bunescu Machine Learning Group Department of Computer Sciences University of Texas at Austin.
Crowd explicit sentiment analysis A. Montejo-Raez, M.C. Diaz-Galiano, F. Martinez-Santiago, L.A. Urena-Lopez Computer Science Department, University of.
1 Generating Comparative Summaries of Contradictory Opinions in Text (CIKM09’)Hyun Duk Kim, ChengXiang Zhai 2010/05/24 Yu-wen,Hsu.
Comparative Experiments on Sentiment Classification for Online Product Reviews Hang Cui, Vibhu Mittal, and Mayur Datar AAAI 2006.
Iterative similarity based adaptation technique for Cross Domain text classification Under: Prof. Amitabha Mukherjee By: Narendra Roy Roll no: Group:
Semantic Compositionality through Recursive Matrix-Vector Spaces
PhD Dissertation Defense Scaling Up Machine Learning Algorithms to Handle Big Data BY KHALIFEH ALJADDA ADVISOR: PROFESSOR JOHN A. MILLER DEC-2014 Computer.
26/01/20161Gianluca Demartini Ranking Categories for Faceted Search Gianluca Demartini L3S Research Seminars Hannover, 09 June 2006.
Reinforcement Learning for Mapping Instructions to Actions S.R.K. Branavan, Harr Chen, Luke S. Zettlemoyer, Regina Barzilay Computer Science and Artificial.
UROP Research Update Citation Function Classification Eric Yulianto A B 22 February 2013.
Concept-Based Analysis of Scientific Literature Chen-Tse Tsai, Gourab Kundu, Dan Roth UIUC.
Semi-Supervised Recognition of Sarcastic Sentences in Twitter and Amazon -Smit Shilu.
A Unified Architecture for Natural Language Processing: Deep Neural Networks with Multitask Learning Ronan Collobert Jason Weston Presented by Jie Peng.
Efficient Estimation of Word Representations in Vector Space By Tomas Mikolov, Kai Chen, Greg Corrado, Jeffrey Dean. Google Inc., Mountain View, CA. Published.
Short Text Similarity with Word Embedding Date: 2016/03/28 Author: Tom Kenter, Maarten de Rijke Source: CIKM’15 Advisor: Jia-Ling Koh Speaker: Chih-Hsuan.
Collaborative Deep Learning for Recommender Systems
Parsing Natural Scenes and Natural Language with Recursive Neural Networks INTERNATIONAL CONFERENCE ON MACHINE LEARNING (ICML 2011) RICHARD SOCHER CLIFF.
Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation EMNLP’14 paper by Kyunghyun Cho, et al.
Data Mining Practical Machine Learning Tools and Techniques
Sentiment analysis using deep learning methods
Jonatas Wehrmann, Willian Becker, Henry E. L. Cagnini, and Rodrigo C
A Deep Memory Network for Chinese Zero Pronoun Resolution
Sentiment analysis algorithms and applications: A survey
CATEGORIZATION OF NEWS ARTICLES USING NEURAL TEXT CATEGORIZER
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Relation Extraction CSCI-GA.2591
A Hierarchical Model of Reviews for Aspect-based Sentiment Analysis
Wei Wei, PhD, Zhanglong Ji, PhD, Lucila Ohno-Machado, MD, PhD
Efficient Estimation of Word Representation in Vector Space
Distributed Representation of Words, Sentences and Paragraphs
Recurrent Neural Networks
INTERSUBJECTIVITY AND SENTIMENT: FROM LANGUAGE TO KNOWLEDGE
Socialized Word Embeddings
Word embeddings Text processing with current NNs requires encoding into vectors. One-hot encoding: N words encoded by length N vectors. A word gets a.
Presentation By: Eryk Helenowski PURE Mentor: Vincent Bindschaedler
Word embeddings (continued)
Sentiment Classification
Presentation transcript:

Duyu Tang, Furu Wei, Nan Yang, Ming Zhou, Ting Liu, Bing Qin Learning Sentiment-Specific Word Embedding for Twitter Sentiment Classification Duyu Tang, Furu Wei, Nan Yang, Ming Zhou, Ting Liu, Bing Qin Harbin Institute of Technology Microsoft Research ACL 2014 報告者:陳思澄 日期:2015/01/09

Outline Introduction Sentiment-Specific Word Embedding Experiments -C & W Model - SSWE Model Experiments -Twitter Sentiment Classification -Word Similarity of Sentiment Lexicons Conclusion

Introduction: Twitter sentiment classification has attracted increasing research interest in recent years. The objective is to classify the sentiment polarity of a tweet as positive, negative or neutral.

For the task of sentiment classification, an effective feature learning method is to compose the representation of a sentence from the representations of the words or phrases it contains. Accordingly, it is a crucial step to learn the word representation (or word embedding), which is a dense, low-dimensional and real-valued vector for a word. Although existing word embedding learning algorithms are intuitive choices, they are not effective enough if directly used for sentiment classification.

Most existing algorithms for learning continuous word representations typically only model the syntactic context of words but ignore the sentiment of text. This is problematic for sentiment analysis as they usually map words with similar syntactic context but opposite sentiment polarity, such as good and bad, to neighboring word vectors.

In this paper, we propose learning sentiment specific word embedding (SSWE) for sentiment analysis. We encode the sentiment information into the continuous representation of words, so that it is able to separate good and bad to opposite ends of the spectrum. We extend the existing word embedding learning algorithm and develop three neural networks to effectively incorporate the supervision from sentiment polarity of text in their loss functions. We learn the sentiment-specific word embedding from tweets, leveraging massive tweets with emoticons as distant-supervised corpora without any manual annotations. (Distant-supervised: Leverage massive noisy-labeled tweets selected by positive and negative emoticons as training set and build sentiment classifiers directly)

C & W Model (Collobert et al.,2011) Collobert et al. (2011) introduce C&W model to learn word embedding based on the syntactic contexts of words. Lm score Lookup table of word embedding w1.

Sentiment-Specific Word Embedding (SSWE)

Basic Model 1 (SSWEh): We utilize the continuous vector of top layer to predict the sentiment distribution of text. Assuming there are K labels, we modify the dimension of top layer in C&W model as K and add a softmax layer upon the top layer. Softmax layer is suitable for this scenario because its outputs are interpreted as conditional probabilities.

Basic Model 2 (SSWEr): The distribution of [0.7,0.3] can also be interpreted as a positive label because the positive score is larger than the negative score. Compared with SSWEh, the softmax layer is removed because SSWEr does not require probabilistic interpretation. is an indicator function reflecting the sentiment polarity of a sentence. Similar with SSWEh, SSWEr also does not generate the corrupted ngram. Positive Score Negative Score

Unified Model (SSWEu): The C&W model learns word embedding by modeling syntactic contexts of words but ignoring sentiment information. By contrast, SSWEh and SSWEr learn sentiment-specific word embedding by integrating the sentiment polarity of sentences but leaving out the syntactic contexts of words. We develop a unified model (SSWEu) in this part, which captures the sentiment information of sentences as well as the syntactic contexts of words.

Experiments:

Twitter Sentiment Classification:

DistSuper: We use the 10 million tweets selected by positive and negative emoticons as training data. NBSVM: trades-off between Naive Bayes and NB-enhanced SVM. RAE: Recursive Autoencoder NRC: NRC builds the top- performed system in SemEval 2013 Twitter sentiment classification track which incorporates diverse sentiment lexicons and many manually designed features.

Conclusion: In this paper, we propose learning continuous word representations as features for Twitter sentiment classification under a supervised learning framework. We learn sentiment-specific word embedding (SSWE) by integrating the sentiment information into the loss functions of three neural network The effectiveness of SSWE has been verified in Twitter sentiment classification and word similarity judgments.