Presentation is loading. Please wait.

Presentation is loading. Please wait.

1 Unsupervised Semantic Parsing Hoifung Poon and Pedro Domingos EMNLP 2009 Best Paper Award Speaker: Hao Xiong.

Similar presentations


Presentation on theme: "1 Unsupervised Semantic Parsing Hoifung Poon and Pedro Domingos EMNLP 2009 Best Paper Award Speaker: Hao Xiong."— Presentation transcript:

1 1 Unsupervised Semantic Parsing Hoifung Poon and Pedro Domingos EMNLP 2009 Best Paper Award Speaker: Hao Xiong

2 2 Outline Motivation Unsupervised semantic parsing Learning and inference Conclusion

3 3 Semantic Parsing Natural language text  Formal and detailed meaning representation (MR) Also called logical form Standard MR language: First-order logic E.g., Microsoft buys Powerset. BUYS(MICROSOFT,POWERSET)

4 4 Shallow Semantic Processing Semantic role labeling Given a relation, identify arguments E.g., agent, theme, instrument Information extraction Identify fillers for a fixed relational template E.g., seminar (speaker, location, time) In contrast, semantic parsing is Formal: Supports reasoning and decision making Detailed: Obtains far more information

5 5 Supervised Learning User provides: Target predicates and objects Example sentences with meaning annotation System learns grammar and produces parser Examples: Zelle & Mooney [1993] Zettlemoyer & Collins [2005, 2007, 2009] Wong & Mooney [2007] Lu et al. [2008] Ge & Mooney [2009]

6 6 Limitations of Supervised Approaches Applicable to restricted domains only For general text Not clear what predicates and objects to use Hard to produce consistent meaning annotation Crucial to develop unsupervised methods Also, often learn both syntax and semantics Fail to leverage advanced syntactic parsers Make semantic parsing harder

7 7 Unsupervised Approaches For shallow semantic tasks, e.g.: Open IE: TextRunner [Banko et al. 2007] Paraphrases: DIRT [Lin & Pantel 2001] Semantic networks: SNE [Kok & Domingos 2008] Show promise of unsupervised methods But … none for semantic parsing

8 8 This Talk: USP First unsupervised approach for semantic parsing Based on Markov Logic [Richardson & Domingos, 2006] Sole input is dependency trees Can be used in general domains Applied it to extract knowledge from biomedical abstracts and answer questions Substantially outperforms TextRunner, DIRT

9 9 Outline Motivation Unsupervised semantic parsing Learning and inference Conclusion

10 10 USP: Key Idea # 1 Target predicates and objects can be learned Viewed as clusters of syntactic or lexical variations of the same meaning BUYS(-,-)   buys, acquires, ’s purchase of, …   Cluster of various expressions for acquisition MICROSOFT   Microsoft, the Redmond software giant, …   Cluster of various mentions of Microsoft

11 11 USP: Key Idea # 2 Relational clustering  Cluster relations with same objects USP  Recursively cluster arbitrary expressions with similar subexpressions Microsoft buys Powerset Microsoft acquires semantic search engine Powerset Powerset is acquired by Microsoft Corporation The Redmond software giant buys Powerset Microsoft’s purchase of Powerset, …

12 12 USP: Key Idea # 2 Relational clustering  Cluster relations with same objects USP  Recursively cluster expressions with similar subexpressions Microsoft buys Powerset Microsoft acquires semantic search engine Powerset Powerset is acquired by Microsoft Corporation The Redmond software giant buys Powerset Microsoft’s purchase of Powerset, … Cluster same forms at the atom level

13 13 USP: Key Idea # 2 Relational clustering  Cluster relations with same objects USP  Recursively cluster expressions with similar subexpressions Microsoft buys Powerset Microsoft acquires semantic search engine Powerset Powerset is acquired by Microsoft Corporation The Redmond software giant buys Powerset Microsoft’s purchase of Powerset, … Cluster forms in composition with same forms

14 14 USP: Key Idea # 2 Relational clustering  Cluster relations with same objects USP  Recursively cluster expressions with similar subexpressions Microsoft buys Powerset Microsoft acquires semantic search engine Powerset Powerset is acquired by Microsoft Corporation The Redmond software giant buys Powerset Microsoft’s purchase of Powerset, … Cluster forms in composition with same forms

15 15 USP: Key Idea # 2 Relational clustering  Cluster relations with same objects USP  Recursively cluster expressions with similar subexpressions Microsoft buys Powerset Microsoft acquires semantic search engine Powerset Powerset is acquired by Microsoft Corporation The Redmond software giant buys Powerset Microsoft’s purchase of Powerset, … Cluster forms in composition with same forms

16 16 USP: Key Idea # 3 Start directly from syntactic analyses Focus on translating them to semantics Leverage rapid progress in syntactic parsing Much easier than learning both

17 17 USP: System Overview Input: Dependency trees for sentences Converts dependency trees into quasi-logical forms (QLFs) QLF subformulas have natural lambda forms Starts with lambda-form clusters at atom level Recursively builds up clusters of larger forms Output: Probability distribution over lambda-form clusters and their composition MAP semantic parses of sentences

18 18 Probabilistic Model for USP Joint probability distribution over a set of QLFs and their semantic parses Use Markov logic A Markov Logic Network (MLN) is a set of pairs ( F i, w i ) where F i is a formula in first-order logic w i is a real number

19 19 Markov Logical Networks undirected graph model log linear model nsubj(n 1,n 2 ) Microsoft(n 2 )buys(n 1 ) Number of true groundings of F i

20 20 Generating Quasi-Logical Forms buys Microsoft Powerset nsubjdobj Convert each node into an unary atom

21 21 Generating Quasi-Logical Forms nsubjdobj n 1, n 2, n 3 are Skolem constants buys(n 1 ) Microsoft(n 2 ) Powerset(n 3 )

22 22 Generating Quasi-Logical Forms nsubjdobj Convert each edge into a binary atom buys(n 1 ) Microsoft(n 2 ) Powerset(n 3 )

23 23 Generating Quasi-Logical Forms Convert each edge into a binary atom buys(n 1 ) Microsoft(n 2 ) Powerset(n 3 ) nsubj(n 1,n 2 )dobj(n 1,n 3 )

24 24 A Semantic Parse buys(n 1 ) Microsoft(n 2 ) Powerset(n 3 ) nsubj(n 1,n 2 )dobj(n 1,n 3 ) Partition QLF into subformulas

25 25 A Semantic Parse buys(n 1 ) Microsoft(n 2 ) Powerset(n 3 ) nsubj(n 1,n 2 )dobj(n 1,n 3 ) Subformula  Lambda form: Replace Skolem constant not in unary atom with a unique lambda variable

26 26 A Semantic Parse buys(n 1 ) Microsoft(n 2 ) Powerset(n 3 ) λx 2.nsubj(n 1,x 2 ) Subformula  Lambda form: Replace Skolem constant not in unary atom with a unique lambda variable λx 3.dobj(n 1,x 3 )

27 27 A Semantic Parse buys(n 1 ) Microsoft(n 2 ) Powerset(n 3 ) λx 2.nsubj(n 1,x 2 ) Follow Davidsonian Semantics Core form: No lambda variable Argument form: One lambda variable λx 3.dobj(n 1,x 3 ) Core form Argument form

28 28 A Semantic Parse buys(n 1 ) Microsoft(n 2 ) Powerset(n 3 ) λx 2.nsubj(n 1,x 2 ) Assign subformula to lambda-form cluster λx 3.dobj(n 1,x 3 )  C BUYS  C MICROSOFT  C POWERSET

29 29 Lambda-Form Cluster buys(n 1 ) Distribution over core forms C BUYS 0.1 acquires(n 1 ) 0.2 …… One formula in MLN Learn weights for each pair of cluster and core form

30 30 Lambda-Form Cluster buys(n 1 ) May contain variable number of argument types C BUYS 0.1 acquires(n 1 ) 0.2 …… A BUYER A BOUGHT A PRICE ……

31 31 Argument Type: A BUYER λx 2.nsubj(n 1,x 2 ) Distributions over argument forms, clusters, and number 0.5 0.4 …… C MICROSOFT 0.2 C GOOGLE 0.1 …… None 0.1 One 0.8 …… λx 2.agent(n 1,x 2 ) Three MLN formulas

32 32 Abstract Lambda Form buys(n 1 )  λx 2.nsubj(n 1,x 2 )  λx 3.dobj(n 1,x 3 ) C BUYS (n 1 )  λx 2. A BUYER (n 1,x 2 )  λx 3. A BOUGHT (n 1,x 3 ) Final logical form is obtained via lambda reduction

33 33 Outline Motivation Unsupervised semantic parsing Learning and inference Conclusion

34 34 Learning Observed: Q (QLFs) Hidden: S (semantic parses) Maximizes log-likelihood of observing the QLFs

35 35 Search Operators MERGE(C 1, C 2 ) : Merge clusters C 1, C 2 E.g.:  buys ,  acquires    buys, acquires  COMPOSE(C 1, C 2 ) : Create a new cluster resulting from composing lambda forms in C 1, C 2 E.g.:  Microsoft ,  Corporation    Microsoft Corporation 

36 36 USP-Learn Initialization: Partition  Atoms Greedy step: Evaluate search operations and execute the one with highest gain in log-likelihood Efficient implementation: Inverted index, etc.

37 37 Search Operations Four Operations Examples

38 38 MAP Semantic Parse Goal: Given QLF Q and learned Θ, find semantic parse S to maximize P Θ (Q, S) Again, use greedy search

39 39 Experiments etc…


Download ppt "1 Unsupervised Semantic Parsing Hoifung Poon and Pedro Domingos EMNLP 2009 Best Paper Award Speaker: Hao Xiong."

Similar presentations


Ads by Google