1 Combining Contexts in Lexicon Learning for Semantic Parsing May 25, 2007 NODALIDA 2007, Tartu, Estonia Chris Biemann University of Leipzig Germany Rainer.

Slides:



Advertisements
Similar presentations
TWO STEP EQUATIONS 1. SOLVE FOR X 2. DO THE ADDITION STEP FIRST
Advertisements

A B C D E F G H I J K L M N O P Q R S T U V W X Y Z
You have been given a mission and a code. Use the code to complete the mission and you will save the world from obliteration…
Fundamentals of Probability
Multistage Sampling.
Advanced Piloting Cruise Plot.
1 Automatic Extension of Feature-based Semantic Lexicons via Contextual Features March 10, th Annual Conference of Gfkl, 2005 Chris Biemann University.
Chapter 6 Cost and Choice. Copyright © 2001 Addison Wesley LongmanSlide 6- 2 Figure 6.1 A Simplified Jam-Making Technology.
© 2008 Pearson Addison Wesley. All rights reserved Chapter Seven Costs.
Chapter 1 The Study of Body Function Image PowerPoint
Copyright © 2011, Elsevier Inc. All rights reserved. Chapter 5 Author: Julia Richards and R. Scott Hawley.
Copyright © 2011, Elsevier Inc. All rights reserved. Chapter 6 Author: Julia Richards and R. Scott Hawley.
1 Copyright © 2010, Elsevier Inc. All rights Reserved Fig 2.1 Chapter 2.
By D. Fisher Geometric Transformations. Reflection, Rotation, or Translation 1.
Growing Every Child! The following slides are examples of questions your child will use in the classroom throughout the year. The questions progress from.
FDA/Industry Workshop September, 19, 2003 Johnson & Johnson Pharmaceutical Research and Development L.L.C. 1 Uses and Abuses of (Adaptive) Randomization:
Business Transaction Management Software for Application Coordination 1 Business Processes and Coordination.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
Title Subtitle.
Properties of Real Numbers CommutativeAssociativeDistributive Identity + × Inverse + ×
My Alphabet Book abcdefghijklm nopqrstuvwxyz.
0 - 0.
ALGEBRAIC EXPRESSIONS
DIVIDING INTEGERS 1. IF THE SIGNS ARE THE SAME THE ANSWER IS POSITIVE 2. IF THE SIGNS ARE DIFFERENT THE ANSWER IS NEGATIVE.
MULTIPLYING MONOMIALS TIMES POLYNOMIALS (DISTRIBUTIVE PROPERTY)
ADDING INTEGERS 1. POS. + POS. = POS. 2. NEG. + NEG. = NEG. 3. POS. + NEG. OR NEG. + POS. SUBTRACT TAKE SIGN OF BIGGER ABSOLUTE VALUE.
MULT. INTEGERS 1. IF THE SIGNS ARE THE SAME THE ANSWER IS POSITIVE 2. IF THE SIGNS ARE DIFFERENT THE ANSWER IS NEGATIVE.
FACTORING ax2 + bx + c Think “unfoil” Work down, Show all steps.
Addition Facts
ALGEBRAIC EXPRESSIONS
Year 6 mental test 5 second questions
Around the World AdditionSubtraction MultiplicationDivision AdditionSubtraction MultiplicationDivision.
ZMQS ZMQS
Solve Multi-step Equations
BT Wholesale October Creating your own telephone network WHOLESALE CALLS LINE ASSOCIATED.
NIPRL Chapter 10. Discrete Data Analysis 10.1 Inferences on a Population Proportion 10.2 Comparing Two Population Proportions 10.3 Goodness of Fit Tests.
ABC Technology Project
Copyright © Cengage Learning. All rights reserved. 6 Sets and Counting.
2 |SharePoint Saturday New York City
VOORBLAD.
Factor P 16 8(8-5ab) 4(d² + 4) 3rs(2r – s) 15cd(1 + 2cd) 8(4a² + 3b²)
Squares and Square Root WALK. Solve each problem REVIEW:
© 2012 National Heart Foundation of Australia. Slide 2.
Lets play bingo!!. Calculate: MEAN Calculate: MEDIAN
Sets Sets © 2005 Richard A. Medeiros next Patterns.
Chapter 5 Test Review Sections 5-1 through 5-4.
GG Consulting, LLC I-SUITE. Source: TEA SHARS Frequently asked questions 2.
Addition 1’s to 20.
25 seconds left…...
Equal or Not. Equal or Not
Slippery Slope
Test B, 100 Subtraction Facts
1 Minimally Supervised Morphological Analysis by Multimodal Alignment David Yarowsky and Richard Wicentowski.
Week 1.
We will resume in: 25 Minutes.
©Brooks/Cole, 2001 Chapter 12 Derived Types-- Enumerated, Structure and Union.
A SMALL TRUTH TO MAKE LIFE 100%
A small truth to make life 100%
A SMALL TRUTH TO MAKE LIFE 100%
1 Unit 1 Kinematics Chapter 1 Day
PSSA Preparation.
Chapter 20 The ISLM Model. Copyright © 2007 Pearson Addison-Wesley. All rights reserved Determination of Aggregate Output.
1 PART 1 ILLUSTRATION OF DOCUMENTS  Brief introduction to the documents contained in the envelope  Detailed clarification of the documents content.
How Cells Obtain Energy from Food
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 14 From Randomness to Probability.
User Defined Functions Lesson 1 CS1313 Fall User Defined Functions 1 Outline 1.User Defined Functions 1 Outline 2.Standard Library Not Enough #1.
Presentation transcript:

1 Combining Contexts in Lexicon Learning for Semantic Parsing May 25, 2007 NODALIDA 2007, Tartu, Estonia Chris Biemann University of Leipzig Germany Rainer Osswald FernUniversität Hagen Germany Richard Socher Saarland University Germany

2 Outline Motivation: lexicon extension for semantic parsing The semantic lexicon HaGenLex Binary features and complex sorts Method: bootstrapping via syntactic contexts Results Discussion

3 Motivation Semantic parsing aims at finding a semantic representation for a sentence Semantic parsing needs as a prerequisite semantic features of words. Semantic features are obtained by manually creating lexicon entries (expensive in terms of time and money) Given a certain amount of manually created lexicon entries, it might be possible to train a classifier in order to find more entries Objective is Precision, Recall is secondary

4 HaGenLex: Semantic Lexicon for German complex sort size: 22,700 entries of these: 13,000 nouns, 6,700 verbs WORDSEMANTIC CLASS Aggressivitätnonment-dyn-abs-situation Agonienonment-stat-abs-situation Agrarproduktnat-discrete Ägypterhuman-object Ahnhuman-object Ahndungnonment-dyn-abs-situation Ähnlichkeitrelation Airbagnonax-mov-art-discrete Airbusmov-nonanimate-con-potag Airportart-con-geogr Ajatollahhuman-object Akademikerhuman-object Akademisierungnonment-dyn-abs-situation Akkordeonnonax-mov-art-discrete Akkreditierungnonment-dyn-abs-situation Akkuax-mov-art-discrete Akquisitionnonment-dyn-abs-situation Akrobathuman-object...

5 Characteristics of complex sorts in HaGenLex In total, 50 complex sorts for nouns are constructed from allowed combinations of: 16 semantic features (binary), e.g. HUMAN+, ARTIFICIAL- 17 sorts (binary), e.g. concrete, abstract-situation... sort (hierarchy) semantic features complex sorts

6 Application: WOCADI-Parser Welche Bücher von Peter Jackson über Expertensysteme wurden bei Addison-Wesley seit 1985 veröffentlicht?

7 General Methodology Distributional Hypothesis projected on syntactic- semantic contexts for nouns: nouns of similar complex sort are found in similar contexts We use three kinds of context elements Adjective Modifier Verb-Subject (deep) Verb-Object (deep) as assigned by the WOCADI parser for training 33 binary classifiers.

8 Data Corpus: 3,068,945 sentences obtained from the Leipzig Corpora Collection parser coverage: 42% verb-deep-subject relations: 430,916 verb-deep-object relations: 408,699 adjective-noun relations: 450,184 Lexicon 11,100 noun entries lexicon extension: 10-fold cross validation on known nouns Also unknown nouns will be classified

9 Algorithm: Initialize the training set; As long as new nouns get classified { calculate class probabilities for each context element; for all yet unclassified nouns n { Multiply class probs of context elements class-wise; Assign the class with highest probabilities to noun n; } Class probabilities per context element: a) count number of per class b) normalize on total number of class wrt. noun classes c) normalize to row sum=1 A threshold regulates the minimum number of different context elements a noun co-occurs with in order to be classified Bootstrapping Mechanism

10 From binary classes to complex sorts Binary classifiers for single features for all three context element types are combined into one feature assignment: –Lenient: voting –Strict: all classifiers for different context types agree Combining the outcome: safe choices ANIMAL +/- ANIMATE +/- ARTIF +/- AXIAL +/-... (16 features)... (17 sorts) ab +/- abs +/- ad +/- as +/- Selection: compatible complex sorts that are minimal w.r.t hierarchy and unambiguous. result class or reject

11 Results: binary classes for different context types =5 =1 most of the binary features are highly biased

12 Combination of context types =1

13 Results for complex sorts =5 =1 Complex sorts with highest training frequency

14 Typical mistakes Pflanze (plant) animal-object instead of plant-object zart, fleischfressend, fressend, verändert, genmanipuliert, transgen, exotisch, selten, giftig, stinkend, wachsend... Nachwuchs (offspring) human-object instead of animal-object wissenschaftlich, qualifiziert, akademisch, eigen, talentiert, weiblich, hoffnungsvoll, geeignet, begabt, journalistisch... Café (café) art-con-geogr instead of nonmov-art-discrete (cf. Restaurant) Wiener, klein, türkisch, kurdisch, romanisch, cyber, philosophisch, besucht, traditionsreich, schnieke, gutbesucht,... Neger (negro) animal-object instead of human-object weiß, dreckig, gefangen, faul, alt, schwarz, nackt, lieb, gut, brav but: Skinhead (skinhead) human-object (ok) {16,17,18,19,20,21,22,23,30}ährig, gleichaltrig, zusammengeprügelt, rechtsradikal, brutal In most cases the wrong class is semantically close. Evaluation metrics did not account for that.

15 Discussion of Results Binary features: Precision >98% for most binary features Assigning the smaller class is hard for bias>0.9 Context types verb-subject and verb-object are better than adjective verb-subject is best single context for complex sorts combination always helps for binary features Complex sorts Todo: more lenient combination procedure to increase recall

16 Conclusion Method for semantic lexicon extension High precision for binary semantic features Unknown nouns: –For 3,755 nouns not in the lexicon, a total of 125,491 binary features was assigned. –For 1,041 unknown nouns, a complex sort was assigned Combination to complex sorts yet to be improved Combination of different context types improves results

17 Any Questions? Thank you very much!