Dan Jurafsky and Chris Potts Lecture 10: Wrap-up CS 424P/ LINGUIST 287 Extracting Social Meaning and Sentiment.

Slides:



Advertisements
Similar presentations
Distant Supervision for Emotion Classification in Twitter posts 1/17.
Advertisements

Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Language and Cognition Colombo, June 2011 Day 8 Aphasia: disorders of comprehension.
Computational Models of Discourse Analysis Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Automatic Prosodic Event Detection Using Acoustic, Lexical, and Syntactic Evidence Sankaranarayanan Ananthakrishnan, Shrikanth S. Narayanan IEEE 2007 Min-Hsuan.
Sentiment Analysis An Overview of Concepts and Selected Techniques.
Methods in Computational Linguistics II Queens College Lecture 1: Introduction.
Presented by Ravi Kiran. Julia Hirschberg Stefan Benus Jason M. Brenier Frank Enos Sarah Friedman Sarah Gilman Cynthia Girand Martin Graciarena Andreas.
Natural Language Processing AI - Weeks 19 & 20 Natural Language Processing Lee McCluskey, room 2/07
An Overview of Text Mining Rebecca Hwa 4/25/2002 References M. Hearst, “Untangling Text Data Mining,” in the Proceedings of the 37 th Annual Meeting of.
Extracting Social Meaning Identifying Interactional Style in Spoken Conversation Jurafsky et al ‘09 Presented by Laura Willson.
On the Correlation between Energy and Pitch Accent in Read English Speech Andrew Rosenberg Weekly Speech Lab Talk 6/27/06.
1 Psych 5500/6500 The t Test for a Single Group Mean (Part 5): Outliers Fall, 2008.
Introduction to machine learning
SI485i : NLP Set 12 Features and Prediction. What is NLP, really? Many of our tasks boil down to finding intelligent features of language. We do lots.
Toshiba Update 04/09/2006 Data-Driven Prosody and Voice Quality Generation for Emotional Speech Zeynep Inanoglu & Steve Young Machine Intelligence Lab.
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2005 Lecture 1 21 July 2005.
Modeling and Finding Abnormal Nodes (chapter 2) 駱宏毅 Hung-Yi Lo Social Network Mining Lab Seminar July 18, 2007.
Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification on Reviews Peter D. Turney Institute for Information Technology National.
TINONS1 Nonlinear SP and Pattern recognition
Skills for evidence-informed practice: Interactive workshop Cambridge 30 April 2009.
Lecture 6: The Ultimate Authorship Problem: Verification for Short Docs Moshe Koppel and Yaron Winter.
Natural Language Processing
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
Concept of Power ture=player_detailpage&v=7yeA7a0u S3A.
Natural Language Processing Introduction. 2 Natural Language Processing We’re going to study what goes into getting computers to perform useful and interesting.
Copyright 2007, Toshiba Corporation. How (not) to Select Your Voice Corpus: Random Selection vs. Phonologically Balanced Tanya Lambert, Norbert Braunschweiler,
 Text Representation & Text Classification for Intelligent Information Retrieval Ning Yu School of Library and Information Science Indiana University.
On Speaker-Specific Prosodic Models for Automatic Dialog Act Segmentation of Multi-Party Meetings Jáchym Kolář 1,2 Elizabeth Shriberg 1,3 Yang Liu 1,4.
1 Statistical NLP: Lecture 9 Word Sense Disambiguation.
Assessment of Morphology & Syntax Expression. Objectives What is MLU Stages of Syntactic Development Examples of Difficulties in Syntax Why preferring.
This work is supported by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior National Business Center contract number.
C ONTEXT AND CULTURE. D O YOU REMEMBER THIS ? Hymes suggests that in order to be able to communicate language, a person should acquire four types of knowledge:
THE NATURE OF TEXTS English Language Yo. Lets Refresh So we tend to get caught up in the themes on English Language that we need to remember our basic.
1 Determining query types by analysing intonation.
1 Statistical NLP: Lecture 7 Collocations. 2 Introduction 4 Collocations are characterized by limited compositionality. 4 Large overlap between the concepts.
HYMES (1964) He developed the concept that culture, language and social context are clearly interrelated and strongly rejected the idea of viewing language.
1 Computation Approaches to Emotional Speech Julia Hirschberg
Automatic Cue-Based Dialogue Act Tagging Discourse & Dialogue CMSC November 3, 2006.
Recognizing Discourse Structure: Speech Discourse & Dialogue CMSC October 11, 2006.
For Monday Read chapter 24, sections 1-3 Homework: –Chapter 23, exercise 8.
For Friday Finish chapter 24 No written homework.
An Unsupervised Approach for the Detection of Outliers in Corpora David Guthrie Louise Guthire, Yorick Wilks The University of Sheffield.
Communicative and Academic English for the EFL Professional.
Combining Speech Attributes for Speech Recognition Jeremy Morris November 9, 2006.
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
Automatic recognition of discourse relations Lecture 3.
1 Running Experiments for Your Term Projects Dana S. Nau CMSC 722, AI Planning University of Maryland Lecture slides for Automated Planning: Theory and.
4. Relationship Extraction Part 4 of Information Extraction Sunita Sarawagi 9/7/2012CS 652, Peter Lindes1.
JOB ADVERT 9 th Meeting. HOW TO ANSWER ANY INTERVIEW QUESTION BY PERRI CAPELLPERRI CAPELL  He suggests when answering job-interview queries applying.
Discovering Relations among Named Entities from Large Corpora Takaaki Hasegawa *, Satoshi Sekine 1, Ralph Grishman 1 ACL 2004 * Cyberspace Laboratories.
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
Extracting and Ranking Product Features in Opinion Documents Lei Zhang #, Bing Liu #, Suk Hwan Lim *, Eamonn O’Brien-Strain * # University of Illinois.
Special Topics in Educational Data Mining HUDK5199 Spring term, 2013 March 6, 2013.
Lexical, Prosodic, and Syntactics Cues for Dialog Acts.
For Monday Read chapter 26 Homework: –Chapter 23, exercises 8 and 9.
Machine Learning in Practice Lecture 9 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Concept-Based Analysis of Scientific Literature Chen-Tse Tsai, Gourab Kundu, Dan Roth UIUC.
Sentiment Analysis Using Common- Sense and Context Information Basant Agarwal 1,2, Namita Mittal 2, Pooja Bansal 2, and Sonal Garg 2 1 Department of Computer.
Lecture 2 Page 1 CS 236 Online Security Policies Security policies describe how a secure system should behave Policy says what should happen, not how you.
King Faisal University جامعة الملك فيصل Deanship of E-Learning and Distance Education عمادة التعلم الإلكتروني والتعليم عن بعد [ ] 1 جامعة الملك فيصل عمادة.
Speech and multimodal Jesse Cirimele. papers “Multimodal interaction” Sharon Oviatt “Designing SpeechActs” Yankelovich et al.
Retrieval of audio testimonials via voice search
Turn-taking and Disfluencies
David Cyphert CS 2310 – Software Engineering
An Overview of Concepts and Selected Techniques
PURE Learning Plan Richard Lee, James Chen,.
Introduction to Sentiment Analysis
From Unstructured Text to StructureD Data
Presentation transcript:

Dan Jurafsky and Chris Potts Lecture 10: Wrap-up CS 424P/ LINGUIST 287 Extracting Social Meaning and Sentiment

General Take-Away Messages Most sentiment papers use really dumb features. The field is young. There’s lots of room for creativity! Consider a broad spectrum of social meaning: such papers are likely to make a valuable contribution. Sentiment is expressed w/words prosody, gesture… The best models will probably synthesize lots of modalities. Look at your data, don't just run a classifier and report F-score.

Better Sentiment Lexicons Build an effective sentiment lexicon and it will get used *a lot*. Witness LIWC, SentiWordNet, Inquirer. Words aren’t enough Sentiment units don't necessarily align with word units; lexicons should contain phrases, constructions, etc. Strings aren’t enough. Neither are (string, category) pairs. We might need the full richness of the context of utterance. Sentiment is domain-dependent

Better Sentiment Lexicons Seed-sets: not necessarily bad, but the field needs to move past hand-built seed-sets, because the nature of the seed-sets is often the most important factor in performance. Graph-based approaches are common, but there is no consensus that they are the best.

Sentiment Classification Sentiment seems to be harder than other classification problems; expect lower scores. Sentiment might be blended and continuous, making classification somewhat inappropriate. Feature presence seems to be better than feature frequency. (The opposite is generally true for other tasks.) Feature selection is important. It is worth finding a measure that maximizes information gain but also considers the strength of the evidence. Sentiment features are highly topic/context dependent.

Prosody Pitch and energy are easy consider range or any other measure of variance in addition to max and min remove outliers; can do this via quartiles, or SD, or dropping 10%, etc A maxim from speech recognition: biggest reduction in error rate always comes from cleaning up training data. Duration/rate of speech rate of speech is easier to compute. pause length, number of pauses, or burstiness (variance of duration features) Jitter/Shimmer seem powerful, but may require cleaner speech

Disfluencies A big untapped feature set Consider the non-"uh/um" ones especially, uh/um don’t seem to have much social meaning Types: restarts, repetitions, you know/i mean/like, word fragments

Flirtation, Dialogue Use Pennebaker or other lexicons only as a first pass to help you build a real features set; Don't rely on the Pennebaker names of classes; look at your data (i.e. at how the lexical cluster behaves). "mean" is very rarely about meaning, nor "like" about preference. Dialogue features more important than lexical ones. build dialogue features from the linguistics literature, or from other labeled dialogue corpora Think about speaker differences (personality, etc). NLP doesn't consider individual speakers nearly as much as it should.

Emotion You need to look at other languages if you are trying to make a general claim. But this is hard. There's very little work on lexical features in emotion detection. Most people in this field seem to be from prosodic backgrounds. So there’s an opening here for exciting research in lexical/dialogue features.

Deception Deception features are weak. Deception features are varied: lexical, intonational, physical. Commonly used domain-independent features: 1st person pronouns, positive and negative words, exclusive terms, ``motion verbs'’. Domain-dependent features are valuable, and the nature of the domain can affect the above profoundly. No single feature identifies deception. Perhaps clusters of them can (in context).

Medical It's not good enough to build a classifier that can find a disease or state if it can’t distinguish it from different diseases/states (i.e. a drunk detector is useless if it can't tell drunk from stressed) Lesson from Agatha Christie detector: many features are genre-specific, and even if the feature isn't, the threshold might be. “I" and "we" seem very useful but it's even better if you look at the data and convince yourself they worked for the right reason.

Politics Bias != subjectivity. Bias is in the eye of the beholder/classifier. Consider the observer's role. Some people are deceptive about their partisanship, but our algorithms have a chance to see through the subterfuge. Classification models benefit from social, relational information. Work in this area can inform journalism and public- policy debates.

Some ideas for future projects Detect power structures from conversation (Enron?) Paranoia detection, compulsiveness detection, autism spectrum, Aspergers. perhaps build a corpus from online forums, autobiography by people with known diagnosis, etc Determine from chat how close friends people are. Tone-checking in Build an effective sentiment lexicon. Perhaps begin with small high-precision seeds (LIWC, GI, etc), and then extend/adapt it in particular domains Mine interesting tasks from Scherer's 5 kinds of affective speech. Use disfluencies to build better drunk-detector

Some ideas for future projects feature presence versus feature counts a function of data size? a function of genre? synonymy? “how helpful is this review”. does it correlate with large counts? using non-text structure in your data (e.g., graph structure) to cluster to do unsupervised lexicon generation humor detection, humor generation better features to describe the sentiment context of words compositionality of vector models of semantics politeness vocatives for addressees in web search with sentiment controversial topic detection blocked wikipedia wikitrust addon