Presentation is loading. Please wait.

Presentation is loading. Please wait.

Template-Based Event Extraction Kevin Reschke – Aug 15 th 2013 Martin Jankowiak, Mihai Surdeanu, Dan Jurafsky, Christopher Manning.

Similar presentations


Presentation on theme: "Template-Based Event Extraction Kevin Reschke – Aug 15 th 2013 Martin Jankowiak, Mihai Surdeanu, Dan Jurafsky, Christopher Manning."— Presentation transcript:

1 Template-Based Event Extraction Kevin Reschke – Aug 15 th 2013 Martin Jankowiak, Mihai Surdeanu, Dan Jurafsky, Christopher Manning

2 Outline Recap from last time Distant supervision Plane crash dataset Current work Fully supervised setting MUC4 terrorism dataset 2 Underlying theme: Joint Inference Models

3 Goal: Knowledge Base Population “… Delta Flight 14 crashed in Mississippi killing 40 … ” … … News Corpus Knowledge Base

4 Distant Supervision Use known events to automatically label training data. Training Knowledge-Base One year after [USAir] Operator [Flight 11] FlightNumber crashed in [Toronto] CrashSite, families of the [200] Fatalities victims attended a memorial service in [Vancouver] NIL.

5 Plane Crash Dataset 80 plane crashes from Wikipedia infoboxes. Training set: 32; Dev set: 8; Test set: 40 Corpus: Newswire data from 1989 – present.

6 Extraction Models Local Model Train and classify each mention independently. Pipeline Model Classify sequentially; use previous label as feature. Captures dependencies between labels. E.g., Passengers and Crew go together: “4 crew and 200 passengers were on board.” Joint Model Searn Algorithm (Daumé III et al., 2009). Jointly models all mentions in a sentence.

7 Results

8 Fully Supervised Setting: MUC4 Terrorism Dataset 4 th Message Understanding Conference (1992). Terrorist activities in Latin America. 1700 docs ( train / dev / test = 1300 / 200 / 200 ). 50/50 mix of relevant and irrelevant doc. 8

9 MUC4 Task 5 slots types: Perpetrator Individual(PerpInd) Perpetrator Organization(PerpOrg) Physical Target(Target) Victim(Victim) Weapon(Weapon) Task: Identify all slot fills in each document. Don’t worry about differentiating multiple events. 9

10 MUC4 Example 10 THE ARCE BATTALION COMMAND HAS REPORTED THAT ABOUT 50 PEASANTS OF VARIOUS AGES HAVE BEEN KIDNAPPED BY TERRORISTS OF THE FARABUNDO MARTI NATIONAL LIBERATION FRONT [FMLN] IN SAN MIGUEL DEPARTMENT. Victim PerpInd PerpOrg

11 MUC4 Example 11 THE ARCE BATTALION COMMAND HAS REPORTED THAT ABOUT 50 PEASANTS OF VARIOUS AGES HAVE BEEN KIDNAPPED BY TERRORISTS OF THE FARABUNDO MARTI NATIONAL LIBERATION FRONT [FMLN] IN SAN MIGUEL DEPARTMENT. PerpInd PerpOrg NIL Victim

12 Baseline Results Local Mention Model Multiclass logistic regression. Pipeline Mention Model Previous non-NIL label (or “none”) is feature for current mention. 12

13 Observation 1: Local context is insufficient. Need sentence-level measure. (Patwardhan & Riloff, 2009) 13 Two bridges were destroyed... in Baghdad last night in a resurgence of bomb attacks in the capital city.... and $50 million in damage was caused by a hurricane that hit Miami on Friday.... to make way for modern, safer bridges that will be constructed early next year. ✓ ✗ ✗

14 Baseline Models + Sentence Relevance Binary relevance classifier – unigram / bigram features HardSent: Discard all mentions in irrelevant sentences. SoftSent: Sentence relevance is feature for mention classification. 14

15 Observation 2: Sentence relevance depends on surrounding context. (Huang & Riloff, 2012) 15 “Obama was attacked.” (political attack vs. terrorist attack) “He use a gun.” (weapon in terrorist event?)

16 Joint Inference Models Idea: Model sentence relevance and mention labels jointly – yield globally optimal decisions. Machinery: Conditional Random Fields (CRFs). Model joint probability of relevance labels and mention labels conditioned on input features. Encode dependencies among labels. Software: Factorie (http://factorie.cs.umass.edu) Flexibly design CRF graph structures. Learning / Classification algorithms with exact and approximate inference. 16

17 First Pass Fully joint model. S M M M Approximate inference a likely culprit. 17

18 Second Pass Two linear-chain CRFs with relevance threshold. S S S M M M 18

19 Analysis Many errors are reasonable extractions, but come from irrelevant documents. Learned CRF model weights: 19 RelLabel > = -0.071687 RelLabel = 0.716669 RelLabel = -1.688919... RelRel = -0.609790 RelRel = -0.469663 RelRel = -0.634649 RelRel = 0.572855 The kidnappers were accused of kidnapping several businessmen for high sums of Money.

20 Possibilities for improvement Label-specific relevance thresholds. Leverage Coref (Skip Chain CRFs). Incorporate doc-level relevance signal. 20

21 State of the art Huang & Riloff (2012) P / R / F 1 : 0.58 / 0.60 / 0.59 CRF sentence model with local mention classifiers. Textual cohesion features to model sentence chains. Multiple binary mention classifiers (SVMs). 21

22 Future Work Apply CRF models to plane crash dataset. New terrorism dataset from Wikipedia. Hybrid models: combine supervised MUC4 data with distant supervision on Wikipedia data. 22

23 Thanks! 23


Download ppt "Template-Based Event Extraction Kevin Reschke – Aug 15 th 2013 Martin Jankowiak, Mihai Surdeanu, Dan Jurafsky, Christopher Manning."

Similar presentations


Ads by Google