Event Extraction: Learning from Corpora Prepared by Ralph Grishman Based on research and slides by Roman Yangarber NYU.

Slides:



Advertisements
Similar presentations
Overview of the TAC2013 Knowledge Base Population Evaluation: English Slot Filling Mihai Surdeanu with a lot help from: Hoa Dang, Joe Ellis, Heng Ji, and.
Advertisements

The Impact of Task and Corpus on Event Extraction Systems Ralph Grishman New York University Malta, May 2010 NYU.
NYU ANLP-00 1 Automatic Discovery of Scenario-Level Patterns for Information Extraction Roman Yangarber Ralph Grishman Pasi Tapanainen Silja Huttunen.
Probabilistic Detection of Context-Sensitive Spelling Errors Johnny Bigert Royal Institute of Technology, Sweden
Automatic Discovery of Useful Facet Terms Wisam Dakka – Columbia University Rishabh Dayal – Columbia University Panagiotis G. Ipeirotis – NYU.
A Self Learning Universal Concept Spotter By Tomek Strzalkowski and Jin Wang Original slides by Iman Sen Edited by Ralph Grishman.
Comparing Methods to Improve Information Extraction System using Subjectivity Analysis Prepared by: Heena Waghwani Guided by: Dr. M. B. Chandak.
Ang Sun Ralph Grishman Wei Xu Bonan Min November 15, 2011 TAC 2011 Workshop Gaithersburg, Maryland USA.
Annotation Free Information Extraction Chia-Hui Chang Department of Computer Science & Information Engineering National Central University
Basic IR: Queries Query is statement of user’s information need. Index is designed to map queries to likely to be relevant documents. Query type, content,
Semi-supervised learning and self-training LING 572 Fei Xia 02/14/06.
CS4705.  Idea: ‘extract’ or tag particular types of information from arbitrary text or transcribed speech.
July 9, 2003ACL An Improved Pattern Model for Automatic IE Pattern Acquisition Kiyoshi Sudo Satoshi Sekine Ralph Grishman New York University.
An Overview of Text Mining Rebecca Hwa 4/25/2002 References M. Hearst, “Untangling Text Data Mining,” in the Proceedings of the 37 th Annual Meeting of.
Inducing Information Extraction Systems for New Languages via Cross-Language Projection Ellen Riloff University of Utah Charles Schafer, David Yarowksy.
Basi di dati distribuite Prof. M.T. PAZIENZA a.a
1 Noun Homograph Disambiguation Using Local Context in Large Text Corpora Marti A. Hearst Presented by: Heng Ji Mar. 29, 2004.
Using Information Extraction for Question Answering Done by Rani Qumsiyeh.
8/13/2004NYCNLP (COLING 2004) Cross-lingual Information Extraction System Evaluation Kiyoshi Sudo Satoshi Sekine Ralph Grishman New York University.
Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang National Central University
Learning Subjective Adjectives from Corpora Janyce M. Wiebe Presenter: Gabriel Nicolae.
INEX 2003, Germany Searching in an XML Corpus Using Content and Structure INEX 2003, Germany Yiftah Ben-Aharon, Sara Cohen, Yael Grumbach, Yaron Kanza,
CSE 730 Information Retrieval of Biomedical Data The use of medical lexicon in biomedical IR.
Automatic Acquisition of Lexical Classes and Extraction Patterns for Information Extraction Kiyoshi Sudo Ph.D. Research Proposal New York University Committee:
Mobile Photos April 17, Auto Extraction of Flickr Tags Unstructured text labels Extract structured knowledge Place and event semantics Scale-structure.
Finding Advertising Keywords on Web Pages Scott Wen-tau YihJoshua Goodman Microsoft Research Vitor R. Carvalho Carnegie Mellon University.
Information Extraction with Unlabeled Data Rayid Ghani Joint work with: Rosie Jones (CMU) Tom Mitchell (CMU & WhizBang! Labs) Ellen Riloff (University.
A Pattern Matching Method for Finding Noun and Proper Noun Translations from Noisy Parallel Corpora Benjamin Arai Computer Science and Engineering Department.
Artificial Intelligence Research Centre Program Systems Institute Russian Academy of Science Pereslavl-Zalessky Russia.
Learning Information Extraction Patterns Using WordNet Mark Stevenson and Mark A. Greenwood Natural Language Processing Group University of Sheffield,
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Semantic Grammar CSCI-GA.2590 – Lecture 5B Ralph Grishman NYU.
1 A study on automatically extracted keywords in text categorization Authors:Anette Hulth and Be´ata B. Megyesi From:ACL 2006 Reporter: 陳永祥 Date:2007/10/16.
Natural Language Processing Group Department of Computer Science University of Sheffield, UK Improving Semi-Supervised Acquisition of Relation Extraction.
Reyyan Yeniterzi Weakly-Supervised Discovery of Named Entities Using Web Search Queries Marius Pasca Google CIKM 2007.
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
A Semantic Approach to IE Pattern Induction Mark Stevenson and Mark Greenwood Natural Language Processing Group University of Sheffield, UK.
Scott Duvall, Brett South, Stéphane Meystre A Hands-on Introduction to Natural Language Processing in Healthcare Annotation as a Central Task for Development.
Czech-English Word Alignment Ondřej Bojar Magdalena Prokopová
Querying Text Databases for Efficient Information Extraction Eugene Agichtein Luis Gravano Columbia University.
NLP: An Information Extraction Perspective Ralph Grishman September 2005 NYU.
Information extraction from text Spring 2003, Part 4 Helena Ahonen-Myka.
A Semantic Approach to IE Pattern Induction Mark Stevenson and Mark A. Greenwood Natural Language Processing Group University of Sheffield, UK.
Bootstrapping for Text Learning Tasks Ramya Nagarajan AIML Seminar March 6, 2001.
Bootstrapping Information Extraction with Unlabeled Data Rayid Ghani Accenture Technology Labs Rosie Jones Carnegie Mellon University & Overture (With.
Comparing Information Extraction Pattern Models Mark Stevenson and Mark A. Greenwood Natural Language Processing Group University of Sheffield, UK.
Bag-of-Words Methods for Text Mining CSCI-GA.2590 – Lecture 2A Ralph Grishman NYU.
Improving Named Entity Translation Combining Phonetic and Semantic Similarities Fei Huang, Stephan Vogel, Alex Waibel Language Technologies Institute School.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
UWMS Data Mining Workshop Content Analysis: Automated Summarizing Prof. Marti Hearst SIMS 202, Lecture 16.
Multi-level Bootstrapping for Extracting Parallel Sentence from a Quasi-Comparable Corpus Pascale Fung and Percy Cheung Human Language Technology Center,
Learning Subjective Nouns using Extraction Pattern Bootstrapping Ellen Riloff School of Computing University of Utah Janyce Wiebe, Theresa Wilson Computing.
Discovering Relations among Named Entities from Large Corpora Takaaki Hasegawa *, Satoshi Sekine 1, Ralph Grishman 1 ACL 2004 * Cyberspace Laboratories.
FILTERED RANKING FOR BOOTSTRAPPING IN EVENT EXTRACTION Shasha Liao Ralph York University.
Learning Extraction Patterns for Subjective Expressions 2007/10/09 DataMining Lab 안민영.
Single Document Key phrase Extraction Using Neighborhood Knowledge.
Basic Syntactic Structures of English CSCI-GA.2590 – Lecture 2B Ralph Grishman NYU.
Toward Entity Retrieval over Structured and Text Data Mayssam Sayyadian, Azadeh Shakery, AnHai Doan, ChengXiang Zhai Department of Computer Science University.
NTNU Speech Lab 1 Topic Themes for Multi-Document Summarization Sanda Harabagiu and Finley Lacatusu Language Computer Corporation Presented by Yi-Ting.
Relation Extraction: Rule-based Approaches CSCI-GA.2590 Ralph Grishman NYU.
An Effective Statistical Approach to Blog Post Opinion Retrieval Ben He, Craig Macdonald, Jiyin He, Iadh Ounis (CIKM 2008)
Presented by: Hassan Sayyadi
NYU Coreference CSCI-GA.2591 Ralph Grishman.
Introduction to Information Extraction
Director, Proteus Project Research in Natural Language Processing
Social Knowledge Mining
Applying Key Phrase Extraction to aid Invalidity Search
CS246: Information Retrieval
Natural Language Processing at NYU: the Proteus Project
Presentation transcript:

Event Extraction: Learning from Corpora Prepared by Ralph Grishman Based on research and slides by Roman Yangarber NYU

Finding Patterns How can we collect patterns? Supervised learning –mark information to be extracted in text –collect information and context = specific patterns –generalize patterns Annotation quite expensive Zipfian distribution of patterns means that annotation of consecutive text is inefficient … the same pattern is annotated many times

Unsupervised learning? The intuition: if we collect documents D R relevant to the scenario, patterns relevant to the scenario will occur more frequently in D R than in the language as a whole (cf. sublanguage predicates in Harris’s distributional analysis)

Riloff ‘96 Corpus manually divided into relevant and irrelevant documents Collect patterns around each noun phrase Score patterns by R log F where R = relevance rate = freq in relevant docs / overall freq Select top-ranked patterns These patterns each find one template slot; combining filled slots into templates is a separate task

Extending the Discovery Procedure Finding relevant documents automatically –Yangarber … use patterns to select documents –Sudo … use keywords and IR engine Defining larger patterns (covering several template slots) –Yangarber … clause structures –Nobata; Sudo … larger structures

Automated Extraction Pattern Discovery Goal: find examples / patterns relevant to a given scenario without any corpus tagging (Yangarber ‘00) Method: –identify a few seed patterns for scenario –retrieve documents containing patterns –find subject-verb-object pattern with high frequency in retrieved documents relatively high frequency in retrieved docs vs. other docs –add pattern to seed and repeat

#1: pick seed pattern Seed:

#2: retrieve relevant documents Seed: Fred retired.... Harry was named president. Maki retired.... Yuki was named president. Relevant documents Other documents

#3: pick new pattern Seed: appears in several relevant documents (top-ranked by Riloff metric) Fred retired.... Harry was named president. Maki retired.... Yuki was named president.

#4: add new pattern to pattern set Pattern set: Note: new patterns added with confidence < 1

Experiment Task: Management succession (as MUC-6) Source: Wall Street Journal Training corpus: ~ 6,000 articles Test corpus: –100 documents: MUC-6 formal training –+ 150 documents judged manually

Pre-processing For each document, find and classify names: –{ person | location | organization | …} Parse document –(regularize passive, relative clauses, etc.) For each clause, collect a candidate pattern: tuple: heads of –[ subject verb direct object object/subject complement locative and temporal modifiers … ]

Experiment: two seed patterns v-appoint = { appoint, elect, promote, name } v-resign = { resign, depart, quit, step-down } Run discovery procedure for 80 iterations

Evaluation Look at discovered patterns –new patterns, missed in manual training Document filtering Slot filling

Discovered patterns

Evaluation: Text Filtering How effective are discovered patterns at selecting relevant documents? IR-style documents matching at least one pattern

How effective are patterns within a complete IE system? MUC-style IE on MUC-6 corpora Caveat: filtered / aligned by hand manual–MUC manual–now Evaluation: Slot filling