Presentation is loading. Please wait.

Presentation is loading. Please wait.

Summarization.

Similar presentations


Presentation on theme: "Summarization."— Presentation transcript:

1 Summarization

2 CS276B Web Search and Mining
Automated Text summarization Tutorial — COLING/ACL’98 Lecture 14 Text Mining II (includes slides borrowed from G. Neumann, M. Venkataramani, R. Altman, L. Hirschman, and D. Radev)

3 an exciting challenge... ...put a book on the scanner, turn the dial to ‘2 pages’, and read the result... ...download 1000 documents from the web, send them to the summarizer, and select the best ones by reading the summaries of the clusters... ...forward the Japanese to the summarizer, select ‘1 par’, and skim the translated summary. 3

4 Headline news — informing
4

5 TV-GUIDES — decision making
5

6 Abstracts of papers — time saving
6

7 Graphical maps — orienting
7

8 Textual Directions — planning
8

9 Cliff notes — Laziness support
9

10 Real systems — Money making
10

11 Questions What kinds of summaries do people want?
What are summarizing, abstracting, gisting,...? How sophisticated must summ. systems be? Are statistical techniques sufficient? Or do we need symbolic techniques and deep understanding as well? What milestones would mark quantum leaps in summarization theory and practice? How do we measure summarization quality? 11

12 What is a Summary? Informative summary Indicative summary
Purpose: replace original document Example: executive summary Indicative summary Purpose: support decision: do I want to read original document yes/no? Example: Headline, scientific abstract

13 ‘Genres’ of Summary? Indicative vs. informative Extract vs. abstract
...used for quick categorization vs. content processing. Extract vs. abstract ...lists fragments of text vs. re-phrases content coherently. Generic vs. query-oriented ...provides author’s view vs. reflects user’s interest. Background vs. just-the-news ...assumes reader’s prior knowledge is poor vs. up-to-date. Single-document vs. multi-document source ...based on one text vs. fuses together many texts. 13

14 Examples of Genres Exercise: summarize the following texts for the following readers: text1: Coup Attempt text2: childrens’ story reader1: your friend, who knows nothing about South Africa. reader2: someone who lives in South Africa and knows the political position. reader3: your 4-year-old niece. reader4: the Library of Congress. 14

15 Why Automatic Summarization?
Algorithm for reading in many domains is: read summary decide whether relevant or not if relevant: read whole document Summary is gate-keeper for large number of documents. Information overload Often the summary is all that is read. Human-generated summaries are expensive.

16 Summary Length (Reuters)
Goldstein et al. 1999

17

18 Summarization Algorithms
Keyword summaries Display most significant keywords Easy to do Hard to read, poor representation of content Sentence extraction Extract key sentences Medium hard Summaries often don’t read well Good representation of content Natural language understanding / generation Build knowledge representation of text Generate sentences summarizing content Hard to do well Something between the last two methods?

19 Sentence Extraction Represent each sentence as a feature vector
Compute score based on features Select n highest-ranking sentences Present in order in which they occur in text. Postprocessing to make summary more readable/concise Eliminate redundant sentences Anaphors/pronouns A woman walks. She smokes. Delete subordinate clauses, parentheticals an anaphora ( /əˈnæfərə/) is a type of expression whose reference depends upon another referential element. E.g., in the sentence 'Sally preferred the company of herself', 'herself' is an anaphoric expression in that it is coref If his in John loves his mot After Amy sneezed all over the tuna salad. So what happened? Did Amy throw it down the garbage disposal or serve it on toast to her friends? No complete thought = fragment. Once Adam smashed the spider. her has the same referent as John, it is an anaphoric pronoun. erential with the expression in subject position.

20 Sentence Extraction: Example
Sigir95 paper on summarization by Kupiec, Pedersen, Chen Trainable sentence extraction Proposed algorithm is applied to its own description (the paper)

21 Sentence Extraction: Example

22 Feature Representation
Fixed-phrase feature Certain phrases indicate summary, e.g. “in summary” Paragraph feature Paragraph initial/final more likely to be important. Thematic word feature Repetition is an indicator of importance Uppercase word feature Uppercase often indicates named entities. (Taylor) Sentence length cut-off Summary sentence should be > 5 words.

23 Feature Representation (cont.)
Sentence length cut-off Summary sentences have a minimum length. Fixed-phrase feature True for sentences with indicator phrase “in summary”, “in conclusion” etc. Paragraph feature Paragraph initial/medial/final Thematic word feature Do any of the most frequent content words occur? Uppercase word feature Is uppercase thematic word introduced?

24 Training Hand-label sentences in training set (good/bad summary sentences) Train classifier to distinguish good/bad summary sentences Model used: Naïve Bayes Can rank sentences according to score and show top n to user.

25 Evaluation Compare extracted sentences with sentences in abstracts

26 Evaluation of features
Baseline (choose first n sentences): 24% Overall performance (42-44%) not very good. However, there is more than one good summary.

27 Multi-Document (MD) Summarization
Summarize more than one document Harder but benefit is large (can’t scan 100s of docs) To do well, need to adopt more specific strategy depending on document set. Other components needed for a production system, e.g., manual post-editing. DUC: government sponsored bake-off 200 or 400 word summaries Longer → easier The Document Understanding Conference (DUC)

28 Types of MD Summaries Single event/person tracked over a long time period Elizabeth Taylor’s bout with pneumonia Give extra weight to character/event May need to include outcome (dates!) Multiple events of a similar nature Marathon runners and races More broad brush, ignore dates An issue with related events Gun control Identify key concepts and select sentences accordingly

29 Determine MD Summary Type
First, determine which type of summary to generate Compute all pairwise similarities Very dissimilar articles → multi-event (marathon) Mostly similar articles Is most frequent concept named entity? Yes → single event/person (Taylor) No → issue with related events (gun control)

30 MultiGen Architecture (Columbia)

31 Generation Ordering according to date Intersection Sentence generator
Find concepts that occur repeatedly in a time chunk Sentence generator

32 Processing Selection of good summary sentences
Elimination of redundant sentences Replace anaphors/pronouns with noun phrases they refer to Need coreference resolution Delete non-central parts of sentences

33 Newsblaster (Columbia)

34 Query-Specific Summarization
So far, we’ve look at generic summaries. A generic summary makes no assumption about the reader’s interests. Query-specific summaries are specialized for a single information need, the query. Summarization is much easier if we have a description of what the user wants.

35 Genre Some genres are easy to summarize
Newswire stories Inverted pyramid structure The first n sentences are often the best summary of length n Some genres are hard to summarize Long documents (novels, the bible) Scientific articles? Trainable summarizers are genre-specific.

36 Discussion Correct parsing of document format is critical.
Need to know headings, sequence, etc. Limits of current technology Some good summaries require natural language understanding Example: President Bush’s nominees for ambassadorships Contributors to Bush’s campaign Veteran diplomats Others

37 Coreference Resolution

38 Coreference Two noun phrases referring to the same entity are said to corefer. Example: Transcription from RL95-2 is mediated through an ERE element at the 5-flanking region of the gene. Coreference resolution is important for many text mining tasks: Information extraction Summarization First story detection

39 Types of Coreference Noun phrases: Transcription from RL95-2 … the gene … Pronouns: They induced apoptosis. Possessives: … induces their rapid dissociation … Demonstratives: This gene is responsible for Alzheimer’s

40 Preferences in pronoun interpretation
Recency: John has an Integra. Bill has a legend. Mary likes to drive it. Grammatical role: John went to the Acura dealership with Bill. He bought an Integra. (?) John and Bill went to the Acura dealership. He bought an Integra. Repeated mention: John needed a car to go to his new job. He decided that he wanted something sporty. Bill went to the Acura dealership with him. He bought an Integra.

41 Preferences in pronoun interpretation
Parallelism: Mary went with Sue to the Acura dealership. Sally went with her to the Mazda dealership. ??? Mary went with Sue to the Acura dealership. Sally told her not to buy anything. Verb semantics: John telephoned Bill. He lost his pamphlet on Acuras. John criticized Bill. He lost his pamphlet on Acuras.

42 An algorithm for pronoun resolution
Two steps: discourse model update and pronoun resolution. Salience values are introduced when a noun phrase that evokes a new entity is encountered. Salience factors: set empirically.

43 Salience weights in Lappin and Leass
Sentence recency 100 Subject emphasis 80 Existential emphasis 70 Accusative emphasis 50 Indirect object and oblique complement emphasis 40 Non-adverbial emphasis Head noun emphasis

44 Lappin and Leass (cont’d)
Recency: weights are cut in half after each sentence is processed. Examples: An Acura Integra is parked in the lot. (subject) There is an Acura Integra parked in the lot. (existential predicate nominal) John parked an Acura Integra in the lot. (object) John gave Susan an Acura Integra. (indirect object) In his Acura Integra, John showed Susan his new CD player. (demarcated adverbial PP)

45 Algorithm Collect the potential referents (up to four sentences back).
Remove potential referents that do not agree in number or gender with the pronoun. Remove potential referents that do not pass intrasentential syntactic coreference constraints. Compute the total salience value of the referent by adding any applicable values for role parallelism (+35) or cataphora (-175). Select the referent with the highest salience value. In case of a tie, select the closest referent in terms of string position. ca·taph·o·ra   [kuh-taf-er-uh]  Show IPA noun Grammar .the use of a word or phrase to refer to a following word or group of words, as the use of the phrase asfollows.

46 Observations Lappin & Leass - tested on computer manuals - 86% accuracy on unseen data. Another well known theory is Centering (Grosz, Joshi, Weinstein), which has an additional concept of a “center”. (More of a theoretical model; less empirical confirmation.)

47 Making Sense of it All... To understand summarization, it helps to consider several perspectives simultaneously: 1. Approaches: basic starting point, angle of attack, core focus question(s): psycholinguistics, text linguistics, computation... 2. Paradigms: theoretical stance; methodological preferences: rules, statistics, NLP, Info Retrieval, AI... 3. Methods: the nuts and bolts: modules, algorithms, processing: word frequency, sentence position, concept generalization... 47

48 Psycholinguistic Approach: 2 Studies
Coarse-grained summarization protocols from professional summarizers (Kintsch and van Dijk, 78): Delete material that is trivial or redundant. Use superordinate concepts and actions. Select or invent topic sentence. 552 finely-grained summarization strategies from professional summarizers (Endres-Niggemeyer, 98): Self control: make yourself feel comfortable. Processing: produce a unit as soon as you have enough data. Info organization: use “Discussion” section to check results. Content selection: the table of contents is relevant. 48

49 Computational Approach: Basics
Top-Down: I know what I want! — don’t confuse me with drivel! User needs: only certain types of info System needs: particular criteria of interest, used to focus search Bottom-Up: I’m dead curious: what’s in the text? User needs: anything that’s important System needs: generic importance metrics, used to rate content 49

50 Query-Driven vs. Text-DRIVEN Focus
Top-down: Query-driven focus Criteria of interest encoded as search specs. System uses specs to filter or analyze text portions. Examples: templates with slots with semantic characteristics; termlists of important terms. Bottom-up: Text-driven focus Generic importance metrics encoded as strategies. System applies strategies over rep of whole text. Examples: degree of connectedness in semantic graphs; frequency of occurrence of tokens. 50

51 Bottom-Up, using Info. Retrieval
IR task: Given a query, find the relevant document(s) from a large set of documents. Summ-IR task: Given a query, find the relevant passage(s) from a set of passages (i.e., from one or more documents). xx xxx xxxx x xx xxxx xxx xx xxx xx xxxxx x xxx xx xxx xx x xxx xx xx xxx x xxx xx xxx x xx x xxxx xxxx xx xx xxxx xxx xxx xx xx xxxx x xxx xx x xx xx xxxxx x x xx xxx xxxxxx xxxxxx x x xxxxxxx xx x xxxxxx xxxx xx xx xxxxx xxx xx x xx xxxx xxx xxxx xx xxxxx xxxxx xx xxx x xxxxx xxx Questions: 1. IR techniques work on large volumes of data; can they scale down accurately enough? 2. IR works on words; do abstracts require abstract representations? 51

52 Top-Down, using Info. Extraction
IE task: Given a template and a text, find all the information relevant to each slot of the template and fill it in. Summ-IE task: Given a query, select the best template, fill it in, and generate the contents. Questions: 1. IE works only for very particular templates; can it scale up? 2. What about information that doesn’t fit into any template—is this a generic limitation of IE? xx xxx xxxx x xx xxxx xxx xx xxx xx xxxxx x xxx xx xxx xx x xxx xx xx xxx x xxx xx xxx x xx x xxxx xxxx xxxx xx xx xxxx xxx xxx xx xx xxxx x xxx xx x xx xx xxxxx x x xx xxx xxxxxx xxxxxx x x xxxxxxx xx x xxxxxx xxxx xx xx xxxxx xxx xx x xx xx xxxx xxx xxxx xx xxxxx xxxxx xx xxx x xxxxx xxx Xxxxx: xxxx Xxx: xxxx Xxx: xx xxx Xx: xxxxx x Xx: x xxx xx Xx: xxx x Xxx: xx Xxx: x 52

53 Paradigms: NLP/IE vs. ir/statistics
Approach: try to ‘understand’ text—re-represent content using ‘deeper’ notation; then manipulate that. Need: rules for text analysis and manipulation, at all levels. Strengths: higher quality; supports abstracting. Weaknesses: speed; still needs to scale up to robust open-domain summarization. IR/Statistics: Approach: operate at lexical level—use word frequency, collocation counts, etc. Need: large amounts of text. Strengths: robust; good for query-oriented summaries. Weaknesses: lower quality; inability to manipulate information at abstract levels. 53

54 Toward the Final Answer...
Problem: What if neither IR-like nor IE-like methods work? Solution: semantic analysis of the text (NLP), using adequate knowledge bases that support inference (AI). Word counting Mrs. Coolidge: “What did the preacher preach about?” Coolidge: “Sin.” Mrs. Coolidge: “What did he say?” Coolidge: “He’s against it.” sometimes counting and templates are insufficient, and then you need to do inference to understand. Inference 54

55 The Optimal Solution... Combine strengths of both paradigms…
...use IE/NLP when you have suitable template(s), ...use IR when you don’t… …but how exactly to do it? 55

56 A Summarization Machine
DOC MULTIDOCS QUERY 50% Very Brief Brief Headline 10% 100% Long ABSTRACTS Extract Abstract ? Indicative Informative CASE FRAMES TEMPLATES CORE CONCEPTS CORE EVENTS RELATIONSHIPS CLAUSE FRAGMENTS INDEX TERMS Generic Query-oriented EXTRACTS Background Just the news 56

57 The Modules of the Summarization Machine
MULTIDOC EXTRACTS E X T R A C I O N I N T E R P A O F I L T E R N G G E N R A T I O ABSTRACTS DOC EXTRACTS CASE FRAMES TEMPLATES CORE CONCEPTS CORE EVENTS RELATIONSHIPS CLAUSE FRAGMENTS INDEX TERMS ? EXTRACTS 57

58

59 Summarization methods (& exercise). Topic Extraction. Interpretation.
Generation. 59

60 Overview of Extraction Methods
Position in the text lead method; optimal position policy title/heading method Cue phrases in sentences Word frequencies throughout the text Cohesion: links among words word co-occurrence coreference lexical chains Discourse structure of the text Information Extraction: parsing and analysis 60

61 POSition-based method (1)
Claim: Important sentences occur at the beginning (and/or end) of texts. Lead method: just take first sentence(s)! Experiments: In 85% of 200 individual paragraphs the topic sentences occurred in initial position and in 7% in final position (Baxendale, 58). Only 13% of the paragraphs of contemporary writers start with topic sentences (Donlan, 80). 61

62 Optimum Position Policy (OPP)
Claim: Important sentences are located at positions that are genre-dependent; these positions can be determined automatically through training (Lin and Hovy, 97). Corpus: newspaper articles (ZIFF corpus). Step 1: For each article, determine overlap between sentences and the index terms for the article. Step 2: Determine a partial ordering over the locations where sentences containing important words occur: Optimal Position Policy (OPP) 62

63 Title-Based Method (1) Claim: Words in titles and headings are positively relevant to summarization. Shown to be statistically valid at 99% level of significance (Edmundson, 68). Empirically shown to be useful in summarization systems. 63

64 Cue-Phrase method (1) Claim 1: Important sentences contain ‘bonus phrases’, such as significantly, In this paper we show, and In conclusion, while non-important sentences contain ‘stigma phrases’ such as hardly and impossible. Claim 2: These phrases can be detected automatically (Kupiec et al. 95; Teufel and Moens 97). Method: Add to sentence score if it contains a bonus phrase, penalize if it contains a stigma phrase. 64

65 Word-frequency-based method (1)
Claim: Important sentences contain words that occur “somewhat” frequently. Method: Increase sentence score for each frequent word. Evaluation: Straightforward approach empirically shown to be mostly detrimental in summarization systems. The resolving power of words words (Luhn, 59) 65

66 Cohesion-based methods
Claim: Important sentences/paragraphs are the highest connected entities in more or less elaborate semantic structures. Classes of approaches word co-occurrences; local salience and grammatical relations; co-reference; lexical similarity (WordNet, lexical chains); combinations of the above. 66

67 Cohesion: WORD co-occurrence (1)
Apply IR methods at the document level: texts are collections of paragraphs (Salton et al., 94; Mitra et al., 97; Buckley and Cardie, 97): Use a traditional, IR-based, word similarity measure to determine for each paragraph Pi the set Si of paragraphs that Pi is related to. Method: determine relatedness score Si for each paragraph, extract paragraphs with largest Si scores. P1 P2 P3 P4 P5 P6 P7 P8 P9 67

68 Word co-occurrence method (2)
In the context of query-based summarization Cornell’s Smart-based approach expand original query compare expanded query against paragraphs select top three paragraphs (max 25% of original) that are most similar to the original query (SUMMAC,98): 71.9% F-score for relevance judgment CGI/CMU approach maximize query-relevance while minimizing redundancy with previous information. (SUMMAC,98): 73.4% F-score for relevance judgment 68

69 Cohesion: Local salience Method
Assumes that important phrasal expressions are given by a combination of grammatical, syntactic, and contextual parameters (Boguraev and Kennedy, 97): CNTX: 50 iff the expression is in the current discourse segment SUBJ: iff the expression is a subject EXST: 70 iff the expression is an existential construction ACC: iff the expression is a direct object HEAD: 80 iff the expression is not contained in another phrase ARG: iff the expression is not contained in an adjunct 69

70 Cohesion: Lexical chains method (1)
Based on (Morris and Hirst, 91) But Mr. Kenny’s move speeded up work on a machine which uses micro-computers to control the rate at which an anaesthetic is pumped into the blood of patients undergoing surgery. Such machines are nothing new. But Mr. Kenny’s device uses two personal-computers to achieve much closer monitoring of the pump feeding the anaesthetic into the patient. Extensive testing of the equipment has sufficiently impressed the authorities which regulate medical equipment in Britain, and, so far, four other countries, to make this the first such machine to be licensed for commercial sale to hospitals. 70

71 Lexical chains-based method (2)
Assumes that important sentences are those that are ‘traversed’ by strong chains (Barzilay and Elhadad, 97). Strength(C) = length(C) - #DistinctOccurrences(C) For each chain, choose the first sentence that is traversed by the chain and that uses a representative set of concepts from that chain. 71

72 Cohesion: Coreference method
Build co-reference chains (noun/event identity, part-whole relations) between query and document - In the context of query-based summarization title and document sentences within document Important sentences are those traversed by a large number of chains: a preference is imposed on chains (query > title > doc) Evaluation: 67% F-score for relevance (SUMMAC, 98) (Baldwin and Morton, 98) 72

73 Cohesion: Connectedness method (1)
(Mani and Bloedorn, 97) Map texts into graphs: The nodes of the graph are the words of the text. Arcs represent adjacency, grammatical, co-reference, and lexical similarity-based relations. Associate importance scores to words (and sentences) by applying the tf.idf metric. Assume that important words/sentences are those with the highest scores. 73

74 Connectedness method (2)
In the context of query-based summarization When a query is given, by applying a spreading-activation algorithms, weights can be adjusted; as a results, one can obtain query-sensitive summaries. Evaluation (Mani and Bloedorn, 97): IR categorization task: close to full-document categorization results. 74

75 Discourse-based method
Claim: The multi-sentence coherence structure of a text can be constructed, and the ‘centrality’ of the textual units in this structure reflects their importance. Tree-like representation of texts in the style of Rhetorical Structure Theory (Mann and Thompson,88). Use the discourse representation in order to determine the most important textual units. Attempts: (Ono et al., 94) for Japanese. (Marcu, 97) for English. 75

76 Rhetorical parsing (Marcu,97)
[With its distant orbit {– 50 percent farther from the sun than Earth –} and slim atmospheric blanket,1] [Mars experiences frigid weather conditions.2] [Surface temperatures typically average about –60 degrees Celsius (–76 degrees Fahrenheit) at the equator and can dip to –123 degrees C near the poles.3] [Only the midday sun at tropical latitudes is warm enough to thaw ice on occasion,4] [but any liquid water formed that way would evaporate almost instantly5] [because of the low atmospheric pressure.6] [Although the atmosphere holds a small amount of water, and water-ice clouds sometimes develop,7] [most Martian weather involves blowing dust or carbon dioxide.8] [Each winter, for example, a blizzard of frozen carbon dioxide rages over one pole, and a few meters of this dry-ice snow accumulate as previously frozen carbon dioxide evaporates from the opposite polar cap.9] [Yet even on the summer pole, {where the sun remains in the sky all day long,} temperatures never warm enough to melt frozen water.10] 76

77 Rhetorical parsing (2) Use discourse markers to hypothesize rhetorical relations rhet_rel(CONTRAST, 4, 5)  rhet_rel(CONTRAT, 4, 6) rhet_rel(EXAMPLE, 9, [7,8])  rhet_rel(EXAMPLE, 10, [7,8]) Use semantic similarity to hypothesize rhetorical relations if similar(u1,u2) then rhet_rel(ELABORATION, u2, u1)  rhet_rel(BACKGROUND, u1,u2) else rhet_rel(JOIN, u1, u2) rhet_rel(JOIN, 3, [1,2])  rhet_rel(ELABORATION, [4,6], [1,2]) Use the hypotheses in order to derive a valid discourse representation of the original text. 77

78 Summarization = selection of the
Rhetorical parsing (3) 2 Elaboration 2 Elaboration 8 Example 2 Background Justification 3 Elaboration 8 Concession 10 Antithesis 1 2 3 4 5 Contrast 7 8 9 10 Summarization = selection of the most important units 2 > 8 > 3, 10 > 1, 4, 5, 7, 9 > 6 4 5 Evidence Cause 5 6 78

79 Information extraction Method (1)
Idea: content selection using templates Predefine a template, whose slots specify what is of interest. Use a canonical IE system to extract from a (set of) document(s) the relevant information; fill the template. Generate the content of the template as the summary. Previous IE work: FRUMP (DeJong, 78): ‘sketchy scripts’ of terrorism, natural disasters, political visits... (Mauldin, 91): templates for conceptual IR. (Rau and Jacobs, 91): templates for business. (McKeown and Radev, 95): templates for news. 79

80 Information Extraction method (2)
Example template: MESSAGE:ID TSL-COL-0001 SECSOURCE:SOURCE Reuters SECSOURCE:DATE 26 Feb 93 Early afternoon INCIDENT:DATE 26 Feb 93 INCIDENT:LOCATION World Trade Center INCIDENT:TYPE Bombing HUM TGT:NUMBER AT LEAST 5 80

81 IE State of the Art MUC conferences (1988–97):
Test IE systems on series of domains: Navy sub-language (89), terrorism (92), business (96),... Create increasingly complex templates. Evaluate systems, using two measures: Recall (how many slots did the system actually fill, out of the total number it should have filled?). Precision (how correct were the slots that it filled?). 81

82 Review of Methods Bottom-up methods Top-down methods
Text location: title, position Cue phrases Word frequencies Internal text cohesion: word co-occurrences local salience co-reference of names, objects lexical similarity semantic rep/graph centrality Discourse structure centrality Information extraction templates Query-driven extraction: query expansion lists co-reference with query names lexical similarity to query 82


Download ppt "Summarization."

Similar presentations


Ads by Google