Semantic Role Labeling. Introduction Semantic Role Labeling AgentThemePredicateLocation.

Slides:



Advertisements
Similar presentations
School of something FACULTY OF OTHER School of Computing FACULTY OF ENGINEERING Chunking: Shallow Parsing Eric Atwell, Language Research Group.
Advertisements

SEMANTIC ROLE LABELING BY TAGGING SYNTACTIC CHUNKS
Natural Language Processing Lecture 2: Semantics.
CONSTRAINED CONDITIONAL MODELS TUTORIAL Jingyu Chen, Xiao Cheng.
INTRODUCTION TO ARTIFICIAL INTELLIGENCE Massimo Poesio Relation Extraction.
Sequence Classification: Chunking Shallow Processing Techniques for NLP Ling570 November 28, 2011.
Layering Semantics (Putting meaning into trees) Treebank Workshop Martha Palmer April 26, 2007.
Natural Language Processing Semantic Roles. Semantics Road Map 1.Lexical semantics 2.Disambiguating words Word sense disambiguation Coreference resolution.
Overview of the Hindi-Urdu Treebank Fei Xia University of Washington 7/23/2011.
计算机科学与技术学院 Chinese Semantic Role Labeling with Dependency-driven Constituent Parse Tree Structure Hongling Wang, Bukang Wang Guodong Zhou NLP Lab, School.
Statistical NLP: Lecture 3
Semantic Role Labeling Abdul-Lateef Yussiff
A Joint Model For Semantic Role Labeling Aria Haghighi, Kristina Toutanova, Christopher D. Manning Computer Science Department Stanford University.
Towards Parsing Unrestricted Text into PropBank Predicate- Argument Structures ACL4 Project NCLT Seminar Presentation, 7th June 2006 Conor Cafferkey.
LING NLP 1 Introduction to Computational Linguistics Martha Palmer April 19, 2006.
Steven Schoonover.  What is VerbNet?  Levin Classification  In-depth look at VerbNet  Evolution of VerbNet  What is FrameNet?  Applications.
1 Words and the Lexicon September 10th 2009 Lecture #3.
NLP and Speech Course Review. Morphological Analyzer Lexicon Part-of-Speech (POS) Tagging Grammar Rules Parser thethe – determiner Det NP → Det.
Answer Extraction: Redundancy & Semantics Ling573 NLP Systems and Applications May 24, 2011.
SRL using complete syntactic analysis Mihai Surdeanu and Jordi Turmo TALP Research Center Universitat Politècnica de Catalunya.
Page 1 Generalized Inference with Multiple Semantic Role Labeling Systems Peter Koomen, Vasin Punyakanok, Dan Roth, (Scott) Wen-tau Yih Department of Computer.
Semantic Role Labeling using Maximum Entropy Model Joon-Ho Lim NLP Lab. Korea Univ.
Extracting Opinions, Opinion Holders, and Topics Expressed in Online News Media Text Soo-Min Kim and Eduard Hovy USC Information Sciences Institute 4676.
Automatic Extraction of Opinion Propositions and their Holders Steven Bethard, Hong Yu, Ashley Thornton, Vasileios Hatzivassiloglou and Dan Jurafsky Department.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Syntax Lecture 8: Verb Types 1. Introduction We have seen: – The subject starts off close to the verb, but moves to specifier of IP – The verb starts.
Probabilistic Parsing Reading: Chap 14, Jurafsky & Martin This slide set was adapted from J. Martin, U. Colorado Instructor: Paul Tarau, based on Rada.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
PropBank, VerbNet & SemLink Edward Loper. PropBank 1M words of WSJ annotated with predicate- argument structures for verbs. –The location & type of each.
September 17, : Grammars and Lexicons Lori Levin.
Assessing the Impact of Frame Semantics on Textual Entailment Authors: Aljoscha Burchardt, Marco Pennacchiotti, Stefan Thater, Manfred Pinkal Saarland.
Interpreting Dictionary Definitions Dan Tecuci May 2002.
1 Statistical Parsing Chapter 14 October 2012 Lecture #9.
Natural Language Processing Introduction. 2 Natural Language Processing We’re going to study what goes into getting computers to perform useful and interesting.
AQUAINT Workshop – June 2003 Improved Semantic Role Parsing Kadri Hacioglu, Sameer Pradhan, Valerie Krugler, Steven Bethard, Ashley Thornton, Wayne Ward,
Ideas for 100K Word Data Set for Human and Machine Learning Lori Levin Alon Lavie Jaime Carbonell Language Technologies Institute Carnegie Mellon University.
Linguistic Essentials
Combining Lexical Resources: Mapping Between PropBank and VerbNet Edward Loper,Szu-ting Yi, Martha Palmer September 2006.
Supertagging CMSC Natural Language Processing January 31, 2006.
LING 6520: Comparative Topics in Linguistics (from a computational perspective) Martha Palmer Jan 15,
CPSC 422, Lecture 27Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27 Nov, 16, 2015.
Commonsense Reasoning in and over Natural Language Hugo Liu, Push Singh Media Laboratory of MIT The 8 th International Conference on Knowledge- Based Intelligent.
11 Project, Part 3. Outline Basics of supervised learning using Naïve Bayes (using a simpler example) Features for the project 2.
ARDA Visit 1 Penn Lexical Semantics at Penn: Proposition Bank and VerbNet Martha Palmer, Dan Gildea, Paul Kingsbury, Olga Babko-Malaya, Bert Xue, Karin.
◦ Process of describing the structure of phrases and sentences Chapter 8 - Phrases and sentences: grammar1.
Lecture 19 Word Meanings II Topics Description Logic III Overview of MeaningReadings: Text Chapter 189NLTK book Chapter 10 March 27, 2013 CSCE 771 Natural.
NLP. Introduction to NLP Last week, Min broke the window with a hammer. The window was broken with a hammer by Min last week With a hammer, Min broke.
Chunk Parsing. Also called chunking, light parsing, or partial parsing. Method: Assign some additional structure to input over tagging Used when full.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
Julia Hockenmaier and Mark Steedman.   The currently best single-model statistical parser (Charniak, 1999) achieves Parseval scores of over 89% on the.
Lec. 10.  In this section we explain which constituents of a sentence are minimally required, and why. We first provide an informal discussion and then.
Relation Extraction (RE) via Supervised Classification See: Jurafsky & Martin SLP book, Chapter 22 Exploring Various Knowledge in Relation Extraction.
10/31/00 1 Introduction to Cognitive Science Linguistics Component Topic: Formal Grammars: Generating and Parsing Lecturer: Dr Bodomo.
Natural Language Processing Vasile Rus
COSC 6336: Natural Language Processing
Leonardo Zilio Supervisors: Prof. Dr. Maria José Bocorny Finatto
English Proposition Bank: Status Report
CSC 594 Topics in AI – Natural Language Processing
Statistical NLP: Lecture 3
INAGO Project Automatic Knowledge Base Generation from Text for Interactive Question Answering.
Answer Extraction: Semantics
Improving a Pipeline Architecture for Shallow Discourse Parsing
Lecture 19 Word Meanings II
Probabilistic and Lexicalized Parsing
Automatic Detection of Causal Relations for Question Answering
Linguistic Essentials
Lecture 19 Word Meanings II
CS224N Section 3: Corpora, etc.
Structure of a Lexicon Debasri Chakrabarti 13-May-19.
Progress report on Semantic Role Labeling
Presentation transcript:

Semantic Role Labeling

Introduction

Semantic Role Labeling AgentThemePredicateLocation

Can we figure out that these have the same meaning? XYZ corporation bought the stock. They sold the stock to XYZ corporation. The stock was bought by XYZ corporation. The purchase of the stock by XYZ corporation... The stock purchase by XYZ corporation... 4

A Shallow Semantic Representation: Semantic Roles Predicates (bought, sold, purchase) represent an event semantic roles express the abstract role that arguments of a predicate can take in the event 5 buyerproto-agent agent More specificMore general

Semantic Role Labeling Semantic Roles

Getting to semantic roles Neo-Davidsonian event representation: Sasha broke the window Pat opened the door Subjects of break and open: Breaker and Opener Deep roles specific to each event (breaking, opening) Hard to reason about them for NLU applications like QA 7

Thematic roles Breaker and Opener have something in common! Volitional actors Often animate Direct causal responsibility for their events Thematic roles are a way to capture this semantic commonality between Breakers and Eaters. They are both AGENTS. The BrokenThing and OpenedThing, are THEMES. prototypically inanimate objects affected in some way by the action 8

Thematic roles One of the oldest linguistic models Indian grammarian Panini between the 7th and 4th centuries BCE Modern formulation from Fillmore (1966,1968), Gruber (1965) Fillmore influenced by Lucien Tesnière’s (1959) Éléments de Syntaxe Structurale, the book that introduced dependency grammar Fillmore first referred to roles as actants (Fillmore, 1966) but switched to the term case 9

Thematic roles A typical set: 10

Thematic grid, case frame, θ-grid 11 thematic grid, case frame, θ -grid Break: AGENT, THEME, INSTRUMENT. Example usages of “break” Some realizations:

Diathesis alternations (or verb alternation) Dative alternation: particular semantic classes of verbs, “verbs of future having” (advance, allocate, offer, owe), “send verbs” (forward, hand, mail), “verbs of throwing” (kick, pass, throw), etc. Levin (1993): 47 semantic classes (“Levin classes”) for 3100 English verbs and alternations. In online resource VerbNet. 12 Break: AGENT, INSTRUMENT, or THEME as subject Give: THEME and GOAL in either order

Problems with Thematic Roles Hard to create standard set of roles or formally define them Often roles need to be fragmented to be defined. Levin and Rappaport Hovav (2015): two kinds of INSTRUMENTS intermediary instruments that can appear as subjects The cook opened the jar with the new gadget. The new gadget opened the jar. enabling instruments that cannot Shelly ate the sliced banana with a fork. *The fork ate the sliced banana. 13

Alternatives to thematic roles 1.Fewer roles: generalized semantic roles, defined as prototypes (Dowty 1991) PROTO-AGENT PROTO-PATIENT 2.More roles: Define roles specific to a group of predicates 14 FrameNet PropBank

Semantic Role Labeling The Proposition Bank (PropBank)

PropBank Palmer, Martha, Daniel Gildea, and Paul Kingsbury The Proposition Bank: An Annotated Corpus of Semantic Roles. Computational Linguistics, 31(1):71–106 16

PropBank Roles Proto-Agent Volitional involvement in event or state Sentience (and/or perception) Causes an event or change of state in another participant Movement (relative to position of another participant) Proto-Patient Undergoes change of state Causally affected by another participant Stationary relative to movement of another participant 17 Following Dowty 1991

PropBank Roles Following Dowty 1991 Role definitions determined verb by verb, with respect to the other roles Semantic roles in PropBank are thus verb-sense specific. Each verb sense has numbered argument: Arg0, Arg1, Arg2,… Arg0: PROTO-AGENT Arg1: PROTO-PATIENT Arg2: usually: benefactive, instrument, attribute, or end state Arg3: usually: start point, benefactive, instrument, or attribute Arg4 the end point (Arg2-Arg5 are not really that consistent, causes a problem for labeling) 18

PropBank Frame Files 19

Advantage of a ProbBank Labeling 20 This would allow us to see the commonalities in these 3 sentences:

Modifiers or adjuncts of the predicate: Arg-M 21 ArgM-

PropBanking a Sentence 22 Martha Palmer 2013 A sample parse tree

The same parse tree PropBanked 23 Martha Palmer 2013

Annotated PropBank Data Penn English TreeBank, OntoNotes 5.0. Total ~2 million words Penn Chinese TreeBank Hindi/Urdu PropBank Arabic PropBank Verb Frames Coverage Count of word sense (lexical units) From Martha Palmer 2013 Tutorial

Plus nouns and light verbs 25 Slide from Palmer 2013

Semantic Role Labeling FrameNet

Capturing descriptions of the same event by different nouns/verbs 27

FrameNet Baker et al. 1998, Fillmore et al. 2003, Fillmore and Baker 2009, Ruppenhofer et al Roles in PropBank are specific to a verb Role in FrameNet are specific to a frame: a background knowledge structure that defines a set of frame-specific semantic roles, called frame elements, includes a set of pred cates that use these roles each word evokes a frame and profiles some aspect of the frame 28

The “Change position on a scale” Frame This frame consists of words that indicate the change of an I TEM ’s position on a scale (the A TTRIBUTE ) from a starting point (I NITIAL VALUE ) to an end point (F INAL VALUE ) 29

The “Change position on a scale” Frame 30

31 The “Change position on a scale” Frame

Relation between frames Inherits from: Is Inherited by: Perspective on: Is Perspectivized in: Uses: Is Used by: Subframe of: Has Subframe(s): Precedes: Is Preceded by: Is Inchoative of: Is Causative of: 32

Relation between frames “cause change position on a scale” Is Causative of: Change_position_on_a_scaleChange_position_on_a_scale Adds an agent Role add.v, crank.v, curtail.v, cut.n, cut.v, decrease.v, development.n, diminish.v, double.v, drop.v, enhance.v, growth.n, increase.v, knock down.v, lower.v, move.v, promote.v, push.n, push.v, raise.v, reduce.v, reduction.n, slash.v, step up.v, swell.v 33

Relations between frames 34 Figure from Das et al 2010

Schematic of Frame Semantics 35 Figure from Das et al (2014)

FrameNet Complexity 36 From Das et al. 2010

FrameNet and PropBank representations 37

Semantic Role Labeling Semantic Role Labeling Algorithm

Semantic role labeling (SRL) The task of finding the semantic roles of each argument of each predicate in a sentence. FrameNet versus PropBank: 39

History Semantic roles as a intermediate semantics, used early in machine translation (Wilks, 1973) question-answering (Hendrix et al., 1973) spoken-language understanding (Nash-Webber, 1975) dialogue systems (Bobrow et al., 1977) Early SRL systems Simmons 1973, Marcus 1980: parser followed by hand-written rules for each verb dictionaries with verb-specific case frames (Levin 1977) 40

Why Semantic Role Labeling A useful shallow semantic representation Improves NLP tasks like: question answering Shen and Lapata 2007, Surdeanu et al machine translation Liu and Gildea 2010, Lo et al

A simple modern algorithm 42

How do we decide what is a predicate If we’re just doing PropBank verbs Choose all verbs Possibly removing light verbs (from a list) If we’re doing FrameNet (verbs, nouns, adjectives) Choose every word that was labeled as a target in training data 43

Semantic Role Labeling 44

Features Headword of constituent Examiner Headword POS NNP Voice of the clause Active Subcategorization of pred VP -> VBD NP PP 45 Named Entity type of constit ORGANIZATION First and last words of constit The, Examiner Linear position,clause re: predicate before

Path Features Path in the parse tree from the constituent to the predicate 46

Frequent path features 47 From Palmer, Gildea, Xue 2010

Final feature vector For “The San Francisco Examiner”, Arg0, [issued, NP, Examiner, NNP, active, before, VP  NP PP, ORG, The, Examiner, ] Other features could be used as well sets of n-grams inside the constituent other path features the upward or downward halves whether particular nodes occur in the path 48

3-step version of SRL algorithm 1.Pruning: use simple heuristics to prune unlikely constituents. 2.Identification: a binary classification of each node as an argument to be labeled or a NONE. 3.Classification: a 1-of-N classification of all the constituents that were labeled as arguments by the previous stage 49

Why add Pruning and Identification steps? Algorithm is looking at one predicate at a time Very few of the nodes in the tree could possible be arguments of that one predicate Imbalance between positive samples (constituents that are arguments of predicate) negative samples (constituents that are not arguments of predicate) Imbalanced data can be hard for many classifiers So we prune the very unlikely constituents first, and then use a classifier to get rid of the rest. 50

Pruning heuristics – Xue and Palmer (2004) Add sisters of the predicate, then aunts, then great-aunts, etc But ignoring anything in a coordination structure 51

A common final stage: joint inference The algorithm so far classifies everything locally – each decision about a constituent is made independently of all others But this can’t be right: Lots of global or joint interactions between arguments Constituents in FrameNet and PropBank must be non-overlapping. A local system may incorrectly label two overlapping constituents as arguments PropBank does not allow multiple identical arguments labeling one constituent ARG0 Thus should increase the probability of another being ARG1 52

How to do joint inference Reranking The first stage SRL system produces multiple possible labels for each constituent The second stage classifier the best global label for all constituents Often a classifier that takes all the inputs along with other features (sequences of labels) 53

More complications: FrameNet We need an extra step to find the frame 54 Predicatevector  ExtractFrameFeatures(predicate,parse) Frame  ClassifyFrame(predicate,predicatevector), Frame)

Features for Frame Identification 55 Das et al (2014)

Not just English 56

Not just verbs: NomBank 57 Meyers et al Figure from Jiang and Ng 2006

Additional Issues for nouns Features: Nominalization lexicon (employment  employ) Morphological stem Healthcare, Medicate  care Different positions Most arguments of nominal predicates occur inside the NP Others are introduced by support verbs Especially light verbs “X made an argument”, “Y took a nap” 58

Semantic Role Labeling Conclusion

Semantic Role Labeling A level of shallow semantics for representing events and their participants Intermediate between parses and full semantics Two common architectures, for various languages FrameNet: frame-specific roles PropBank: Proto-roles Current systems extract by parsing sentence Finding predicates in the sentence For each one, classify each parse tree constituent 60