Ani Nenkova Lucy Vanderwende Kathleen McKeown SIGIR 2006.

Slides:



Advertisements
Similar presentations
Debt Sustainability and Debt Composition UNCTAD Paper by Heiner Flassbeck and Ugo Panizza.
Advertisements

1 Text Summarization: News and Beyond Kathleen McKeown Department of Computer Science Columbia University.
Chi Squared Tests. Introduction Two statistical techniques are presented. Both are used to analyze nominal data. –A goodness-of-fit test for a multinomial.
4-4 Multiplication Rule The basic multiplication rule is used for finding P(A and B), the probability that event A occurs in a first trial and event B.
Predicting domain-domain interactions using a parsimony approach Katia Guimaraes, Ph.D. NCBI / NLM / NIH.
Text Specificity and Impact on Quality of News Summaries Annie Louis & Ani Nenkova University of Pennsylvania June 24, 2011.
Database Management Systems, R. Ramakrishnan1 Computing Relevance, Similarity: The Vector Space Model Chapter 27, Part B Based on Larson and Hearst’s slides.
1 Natural Language Processing for the Web Prof. Kathleen McKeown 722 CEPSR, Office Hours: Wed, 1-2; Mon 3-4 TA: Fadi Biadsy 702 CEPSR,
Ensemble Learning: An Introduction
Approaches to automatic summarization Lecture 5. Types of summaries Extracts – Sentences from the original document are displayed together to form a summary.
1 Discussion Class 3 Inverse Document Frequency. 2 Discussion Classes Format: Questions. Ask a member of the class to answer. Provide opportunity for.
Multiplication Rule: Basics
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved. Lecture Slides Elementary Statistics Eleventh Edition and the Triola.
Analyzing Sentiment in a Large Set of Web Data while Accounting for Negation AWIC 2011 Bas Heerschop Erasmus School of Economics Erasmus University Rotterdam.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Section 4-2 Basic Concepts of Probability.
1 POPULATION PROJECTIONS Session 6 - Introduction to population projections Ben Jarabi Population Studies & Research Institute University of Nairobi.
Clustering Unsupervised learning Generating “classes”
Query session guided multi- document summarization THESIS PRESENTATION BY TAL BAUMEL ADVISOR: PROF. MICHAEL ELHADAD.
CHAPTER 5 – Session 11 United Nations Statistics Division/DESA Revision of the Index of Industrial Production (IIP) publication by the UNSD.
Mining and Summarizing Customer Reviews
Systems Analysis – Analyzing Requirements.  Analyzing requirement stage identifies user information needs and new systems requirements  IS dev team.
Office of Institutional Research, Planning and Assessment January 24, 2011 UNDERSTANDING THE DIAGNOSTIC GUIDE.
Adaptive News Access Daniel Billsus Presented by Chirayu Wongchokprasitti.
A Compositional Context Sensitive Multi-document Summarizer: Exploring the Factors That Influence Summarization Ani Nenkova, Stanford University Lucy Vanderwende,
1 Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved. Basic Principle of Statistics: Rare Event Rule If, under a given assumption,
1 Text Summarization: News and Beyond Kathleen McKeown Department of Computer Science Columbia University.
Are Atlantic basin tropical cyclone intensity forecasts improving? Jonathan R. Moskaitis 67 th IHC / 2013 Tropical Cyclone Research Forum Naval Research.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
EASE: An Effective 3-in-1 Keyword Search Method for Unstructured, Semi-structured and Structured Data Cuoliang Li, Beng Chin Ooi, Jianhua Feng, Jianyong.
1 Sentence-extractive automatic speech summarization and evaluation techniques Makoto Hirohata, Yosuke Shinnaka, Koji Iwano, Sadaoki Furui Presented by.
1 Computing Relevance, Similarity: The Vector Space Model.
Understanding the Local Composite Index (LCI) December 5, 2013.
An Asymptotic Analysis of Generative, Discriminative, and Pseudolikelihood Estimators by Percy Liang and Michael Jordan (ICML 2008 ) Presented by Lihan.
CEN st Lecture CEN 4021 Software Engineering II Instructor: Masoud Sadjadi Monitoring (POMA)
Korea Maritime and Ocean University NLP Jung Tae LEE
1 Sentence Extraction-based Presentation Summarization Techniques and Evaluation Metrics Makoto Hirohata, Yousuke Shinnaka, Koji Iwano and Sadaoki Furui.
O PTIMAL SERVICE TASK PARTITION AND DISTRIBUTION IN GRID SYSTEM WITH STAR TOPOLOGY G REGORY L EVITIN, Y UAN -S HUN D AI Adviser: Frank, Yeong-Sung Lin.
DOCUMENT UPDATE SUMMARIZATION USING INCREMENTAL HIERARCHICAL CLUSTERING CIKM’10 (DINGDING WANG, TAO LI) Advisor: Koh, Jia-Ling Presenter: Nonhlanhla Shongwe.
Do Summaries Help? A Task-Based Evaluation of Multi-Document Summarization Kathleen McKeown, Rebecca J. Passonneau David K. Elson, Ani Nenkova, Julia Hirschberg.
UWMS Data Mining Workshop Content Analysis: Automated Summarizing Prof. Marti Hearst SIMS 202, Lecture 16.
Dependence Language Model for Information Retrieval Jianfeng Gao, Jian-Yun Nie, Guangyuan Wu, Guihong Cao, Dependence Language Model for Information Retrieval,
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
A Novel Relational Learning-to- Rank Approach for Topic-focused Multi-Document Summarization Yadong Zhu, Yanyan Lan, Jiafeng Guo, Pan Du, Xueqi Cheng Institute.
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
LexPageRank: Prestige in Multi-Document Text Summarization Gunes Erkan, Dragomir R. Radev (EMNLP 2004)
Univariate Gaussian Case (Cont.)
An evolutionary approach for improving the quality of automatic summaries Constantin Orasan Research Group in Computational Linguistics School of Humanities,
The P YTHY Summarization System: Microsoft Research at DUC 2007 Kristina Toutanova, Chris Brockett, Michael Gamon, Jagadeesh Jagarlamudi, Hisami Suzuki,
Think of a sentence to go with this picture. Can you use any of these words? then if so while though since when Try to use interesting adjectives, powerful.
N-Gram Model Formulas Word sequences Chain rule of probability Bigram approximation N-gram approximation.
Chi-Två Test Kapitel 6. Introduction Two statistical techniques are presented, to analyze nominal data. –A goodness-of-fit test for the multinomial experiment.
SENG521 (Fall SENG 521 Software Reliability & Testing Preparing for Test (Part 6a) Department of Electrical & Computer Engineering,
Using Blog Properties to Improve Retrieval Gilad Mishne (ICWSM 2007)
An Effective Statistical Approach to Blog Post Opinion Retrieval Ben He, Craig Macdonald, Jiyin He, Iadh Ounis (CIKM 2008)
Stats Methods at IC Lecture 3: Regression.
Plan for Today’s Lecture(s)
Section 4-4 Multiplication Rule: Basics.
Hardware & Software Reliability
POPULATION PROJECTIONS
Erasmus University Rotterdam
A Consensus-Based Clustering Method
Effect of Measurement Error on SPC
John Frazier and Jonathan perrier
Gerald Dyer, Jr., MPH October 20, 2016
Contributors Jeremy Brown, Bryan Winters, and Austin Ray
CS224N: Query Focused Multi-Document Summarization
LECTURE 09: BAYESIAN LEARNING
Semantic Similarity Methods in WordNet and their Application to Information Retrieval on the Web Yizhe Ge.
Statistics 5/19/2019.
Bug Localization with Combination of Deep Learning and Information Retrieval A. N. Lam et al. International Conference on Program Comprehension 2017.
Presentation transcript:

Ani Nenkova Lucy Vanderwende Kathleen McKeown SIGIR 2006

 Introduction  Content word frequency  Choice of composition function  Experiment

 In this paper, we study the contribution to summarization of three factors related to frequency: content word frequency  composition functions for estimating sentence importance from word frequency  adjustment of frequency weights based on context.

 The high frequency words from the input are very likely to appear in the human models  For the automatic summarizer, the trend to include more frequent words is preserved, but the numbers are lower than those for the human summaries and the overlap between the machine summary

 the high frequency words in the input will tend to appear in some human model.  we want to partition the words in the input into five classes  high class number is associated with higher frequency in the input for the words in the class.  A word falls in C 0 if it does not appear in any of the human summaries  Formalizing frequency: the multinomial model

 1 : Compute the probability distribution over the words wi appearing in the input  Only verbs, nouns, adjectives and numbers are considered in the computation of the probability distribution.  2 : Assign an importance weight to each sentence

 1: The notion of what is most important to include in the summary changes depending on what information has already been included in the summary.  2 : By updating the probabilities in this intuitive way, we also allow words with initially low probability to have higher impact on the choice of subsequent sentences.  3 : The update of word probability gives a natural way to deal with the redundancy in the multi-document input.  the same unit twice in the same summary is rather improbable.

 is significantly worse than and SUM Avr and is in fact very close to baseline performance.

 All three metrics indicate that the content selection capability of the summarizer is affected by the removal of the context adjustment step.  According to ROUGE-1, removing the context adjustment leads to significantly lower results

 SUMAvr was one of the systems with the lowest amount of repetition in its summaries  These results confirm our intuition that the weight update of words to adjust for context is sufficient for dealing with duplication removal problems.

 When context is taken into account and probabilities are adjusted when the word has already appeared in the summary, performance based on content shows an improvement, but more importantly, repetition in the summary significantly decreases.  These results suggest that the more complex combination of features used by state-of-the- art systems