Computational Models of Discourse Analysis Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.

Slides:



Advertisements
Similar presentations
Close Reading A revision guide to question types.
Advertisements

Relevance Feedback Limitations –Must yield result within at most 3-4 iterations –Users will likely terminate the process sooner –User may get irritated.
Dan Jurafsky Lecture 4: Sarcasm, Alzheimers, +Distributional Semantics Computational Extraction of Social and Interactional Meaning SSLST, Summer 2011.
Huffman code and ID3 Prof. Sin-Min Lee Department of Computer Science.
May 2006CLINT-LN Parsing1 Computational Linguistics Introduction Approaches to Parsing.
Problem Semi supervised sarcasm identification using SASI
Semi Supervised Recognition of Sarcastic Sentences in Twitter and Amazon Dmitry DavidovOren TsurAri Rappoport.
Discussion (20 points) and Final Assignment (200 points)
Bag-of-Words Methods for Text Mining CSCI-GA.2590 – Lecture 2A
Machine Learning in Practice Lecture 3 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Information Retrieval in Practice
Search Engines and Information Retrieval
Information Extraction and Ontology Learning Guided by Web Directory Authors:Martin Kavalec Vojtěch Svátek Presenter: Mark Vickers.
SLIDE 1IS 202 – FALL 2003 Lecture 26: Final Review Prof. Ray Larson & Prof. Marc Davis UC Berkeley SIMS Tuesday and Thursday 10:30 am - 12:00.
Chapter 4 Query Languages.... Introduction Cover different kinds of queries posed to text retrieval systems Keyword-based query languages  include simple.
Automated Essay Evaluation Martin Angert Rachel Drossman.
English GCSE Revision. Section A - Reading There are essentially 5 reading questions as Q1 has two parts. You are being tested on your reading, not your.
Query Rewriting Using Monolingual Statistical Machine Translation Stefan Riezler Yi Liu Google 2010 Association for Computational Linguistics.
Search Engines and Information Retrieval Chapter 1.
APPLICATIONS OF DATA MINING IN INFORMATION RETRIEVAL.
TagHelper: Basics Part 1 Carolyn Penstein Rosé Carnegie Mellon University Funded through the Pittsburgh Science of Learning Center and The Office of Naval.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
Titling Your Essay How to Create a Catchy, but Informative, Title for an Analytical Essay.
Establishing potentially useful evidence  Anecdotal evidence: Based on someone’s experience. Statistical evidence: Based on statistics.
1 Statistical NLP: Lecture 9 Word Sense Disambiguation.
1 University of Palestine Topics In CIS ITBS 3202 Ms. Eman Alajrami 2 nd Semester
AP English Language and Composition Annotations. How and why to annotate? Make a key Three purposes.
Comparing and Ranking Documents Once our search engine has retrieved a set of documents, we may want to Rank them by relevance –Which are the best fit.
Evaluation of (Search) Results How do we know if our results are any good? Evaluating a search engine  Benchmarks  Precision and recall Results summaries:
Close Reading Tips and Tricks. Understanding Questions It is vital that you always use your own words. Only include a quote if you are asked to ‘pick.
Bag-of-Words Methods for Text Mining CSCI-GA.2590 – Lecture 2A Ralph Grishman NYU.
Query Suggestion. n A variety of automatic or semi-automatic query suggestion techniques have been developed  Goal is to improve effectiveness by matching.
Carolyn Penstein Rosé Language Technologies Institute Human-Computer Interaction Institute School of Computer Science With funding from the National Science.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
Automatic recognition of discourse relations Lecture 3.
©2012 Paula Matuszek CSC 9010: Text Mining Applications Lab 3 Dr. Paula Matuszek (610)
Multi-level Bootstrapping for Extracting Parallel Sentence from a Quasi-Comparable Corpus Pascale Fung and Percy Cheung Human Language Technology Center,
Iterative similarity based adaptation technique for Cross Domain text classification Under: Prof. Amitabha Mukherjee By: Narendra Roy Roll no: Group:
Computational Models of Discourse Analysis Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Extending SASI to Satirical Product Reviews: A Preview Bernease Herman University of Michigan Monday, April 22, 2013.
Computational Models of Discourse Analysis Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Computational Models of Discourse Analysis Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
These are the pre-survey results for the edTPA Boot Camp instrument. The first question assessed licensure area. Most respondents were Early Childhood,
Machine Learning in Practice Lecture 10 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Computational Models of Discourse Analysis Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Writing Exercise Try to write a short humor piece. It can be fictional or non-fictional. Essay by David Sedaris.
Machine Learning in Practice Lecture 6 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Machine Learning in Practice Lecture 9 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Semi-Supervised Recognition of Sarcastic Sentences in Twitter and Amazon -Smit Shilu.
Computational Models of Discourse Analysis Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Multi-Class Sentiment Analysis with Clustering and Score Representation Yan Zhu.
Machine Learning in Practice Lecture 9 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Computational Models of Discourse Analysis Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Approaches to Machine Translation
Computational Models of Discourse Analysis
The more difficult topics
Erasmus University Rotterdam
Learning Outcomes Carolynn Rankin YULIS Friday 5th May 2006
Today’s Agenda and essential questions
Q1-Identify and Interpret List four things from the text about…
Statistical NLP: Lecture 9
Approaches to Machine Translation
Machine Learning in Practice Lecture 23
Computational Models of Discourse Analysis
Machine Learning in Practice Lecture 27
Computational Models of Discourse Analysis
Interpreting the Rubric
Introduction to Sentiment Analysis
Statistical NLP : Lecture 9 Word Sense Disambiguation
Presentation transcript:

Computational Models of Discourse Analysis Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute

Warm Up Discussion Look at the analysis I have passed out  Note: inscribed sentiment is underlined and invoked sentiment is italicized and relatively frequent words that appear in either of these types of expressions have been marked in bold Do you see any sarcastic comments here? What if any connection do you see between sentiment and sarcasm? Keeping in mind the style of templates you read about, do you see any snippets of text in these examples that you think would make good templates?

Patterns I see Inscribed sentiment  About 24% of words in underlined segments are relatively high frequency  3 “useful” patterns out of 18 underlined portions  Examples: Like Good More CW than CW Invoked sentiment  About 39% of words were relatively high frequency  About 7 possibly useful patterns out of 17, but only 3 really look unambigous  Examples CW like one CW the CW of the CW Makes little CW to CW to the CW CW and CW of an CW Like CW on CW Leave you CW a little CW and CW CW more like a CW

Unit 3 Plan 3 papers we will discuss all give ideas for using context (at different grain sizes)  Local patterns without syntax Using bootstrapping  Local patterns with syntax Using a parser  Rhetorical patterns within documents Using a statistical modeling technique The first two papers introduce techniques that could feasibly be used in your Unit 3 assignment

Student Comment: Point of Discussion To improve performance language technologies seem to approach the task in either one of two ways. First of all approaches attempt to generate a better abstract model that provides the translation mechanism between a string of terms (sentence) and our human mental model of sentiment in language. Alternatively some start with a baseline and try to find a corpus or dictionary of terms that provides evidence for sentiment. Please clarify

Connection between Appraisal and Sarcasm Student Comment: I’m not exactly sure how one would go about applying appraisal theory to something as elusive as sarcasm. A sarcastic example of invoked negative sentiment from Martin and White, p 72

Inscribed versus Invoked Do we see signposts that tell us how to interpret invoked appraisals?

Overview of Approach Start with small amount of labeled data Generate patterns from examples  Select those that appear in training data more than once and don’t appear both in a 1 and a 5 labeled example Expand data through search using examples from labeled data as queries (take top 50 snippet results) Represent data in terms of templatized patterns Modified kNN classification approach How could you do this with SIDE? 1 Build a feature extractor to generate the set of patterns 2 Use search to set up expanded set of data 3 Apply generated patterns to expanded set of data 4 Use kNN classification

Pattern Generation Classify words into high frequency (HFW) versus content words (CW)  HFWs occur at least 100 times per million words  CW occur no more than 1000 times per million words Also add [product], [company], [title] as additional HFWs Constraints on patterns: 2-6 HFWs, 1-6 slots for CWs, patterns start and end with HFWs Would Appraisal theory suggest other categories of words?

Expand Data: “Great for Insomniacs …” What could they have done instead?

Pattern Selection Approach: Select those that appear in training data more than once and don’t appear both in a 1 and a 5 labeled example Could have used an attribute selection technique like Chi-squared attribute evaluation What do you see as the trade-offs between these approaches?

Representing Data as a Vector Most of the features were from the generated patterns Also included punctuation based features  Number of !, number of ?, number of quotes, number capitalized words What other features would you use? What modifications to feature weights would you propose?

Modified kNN Is there a simpler approach? Weighted average so majority class matches count more.

Evaluation I am …rather wary of the effectiveness of their approach because it seems that they cherry picked a heuristic ‘star- sentiment’ baseline to compare their results to in table 3 but do not offer a similar baseline for table 2. Baseline technique: Count as positive examples those that have a highly negative star rating but lots of positive words. Is this really a strong baseline? Look at the examples from the paper.

Evaluation What do you conclude from this? What surprises you?

Revisit: Overview of Approach Start with small amount of labeled data Generate patterns from examples  Select those that appear in training data more than once and don’t appear both in a 1 and a 5 labeled example Expand data through search using examples from labeled data as queries (take top 50 snippet results) Represent data in terms of templatized patterns Modified kNN classification approach How could you do this with SIDE? 1 Build a feature extractor to generate the set of patterns 2 Use search to set up expanded set of data 3 Apply generated patterns to expanded set of data 4 Use kNN classification

What would it take to achieve inter-rater reliability? You can find definitions and examples on the website, just like in the book, but it’s not enough… Strategies  Simplify – are there distinctions that don’t buy us much anyway?  Add constraints  Identify boarderline cases  Use decision trees

What would it take to achieve inter-rater reliability? Look at Beka and Elijah’s analyses in comparison with mine  What were our big disagreements?  How would we resolve them?

Questions?