Presentation is loading. Please wait.

Presentation is loading. Please wait.

LING 581: Advanced Computational Linguistics Lecture Notes April 27th.

Similar presentations


Presentation on theme: "LING 581: Advanced Computational Linguistics Lecture Notes April 27th."— Presentation transcript:

1 LING 581: Advanced Computational Linguistics Lecture Notes April 27th

2 TCE 2 nd last class today…

3 WordNet Homework If you haven’t already, you should have emailed me this report

4 QA Homework Idea: evaluate the feasibility of QA on the web – using TREC 9 QA examples – programming it up is optional – see and appreciate why it’s hard to do... Steps: – Pick 3 query groups – Simulate (programmatically) QA – use the Collins parser and WordNet to find answers to the queries – submit report (before final class next week) Example Question group What kind of animal was Winnie the Pooh? Winnie the Pooh is what kind of animal? What species was Winnie the Pooh? Winnie the Pooh is an imitation of which animal? What was the species of Winnie the Pooh?

5 Example trees reformulate Qs into declarative sentences with missing wh-phrase –____ (kind of animal) is winnie the pooh –winnie the pooh is ____ (species) –winnie the pooh is an imitation of ____ (animal) –the species of winnie the pooh is ____

6 Example Answers: Winnie the Pooh is such a popular character in Poland Winnie-the-Pooh Is My Co- worker Winnie the Pooh is a little, adorable and cute bear obsessed by honey. Winnie-the-Pooh is so fat. Winnie the Pooh is one of the things most closest to my heart Winnie the Pooh is his usual befuddled self

7 Example Original declarative form: –winnie the pooh is ____ (species) Check semantic relatedness of extracted head words using WordNet: – character – co-worker – little – one – self Here, look at shortest paths Summary HeadwordLength #nodes –bear69258 –character61072 –one7734 –co-worker76488 –self714456 –little1028706 Constraints: length < #nodes

8 Other resources XWN: http://xwn.hlt.utdallas.edu/ glosses in Logical Form

9 XWN Applications – The Extended WordNet may be used as a Core Knowledge Base for applications such as Question Answering, Information Retrieval, Information Extraction, Summarization, Natural Language Generation, Inferences, and other knowledge intensive applications. – The glosses contain a part of the world knowledge since they define the most common concepts of the English language.

10 XWN Example : Dan Moldovan and Adrian Novischi, Lexical Chains for Question Answering, COLING 2002Lexical Chains for Question Answering

11 COALS Take a look at an alternative to WordNet for computing similarity – WordNet: handbuilt system – COALS: the correlated occurrence analogue to lexical semantics (Rohde et al. 2004) a instance of a vector-based statistical model for similarity – e.g., see also Latent Semantic Analysis (LSA) – Singular Valued Decomposition (SVD) » sort by singular values, take top k and reduce the dimensionality of the co- occurrence matrix to rank k based on weighted co-occurrence data from large corpora

12 COALS Basic Idea: – compute co-occurrence counts for (open class) words from a large corpora – corpora: Usenet postings over 1 month 9 million (distinct) articles 1.2 billion word tokens 2.1 million word types – 100,000th word occurred 98 times – co-occurrence counts based on a ramped weighting system with window size 4 – excluding closed-class items 44 wiwi 33 22 11 w i-1 w i-2 w i-3 w i-4 w i+4 w i+3 w i+2 w i+1

13 COALS Example:

14 COALS available online http://dlt4.mit.edu/~dr/COALS/similarity.php

15 Computing Similarity

16 Worked Example: zealous run connectbf/3 –?- connectbf(impassioned,zealous,X). –X = 10 ? –?- connectbf(zealous,impassioned,X). –X = 9 ? compare to b. ravenous –?- connectbf(ravenous,zealous,X). –no –?- connectbf(zealous,ravenous,X). shortest link between impassioned and zealous Old Code: WordNet 1.7.1

17 Worked Example: zealous shortest path between ravenous and zealous

18 Task: Match each word in the first column with its definition in the second column accolade abate aberrant abscond acumen abscission acerbic accretion abjure abrogate deviation abolish keen insight lessen in intensity sour or bitter building up depart secretly renounce removal praise

19 Task: Match each word in the first column with its definition in the second column accolade abate aberrant abscond acumen abscission acerbic accretion abjure abrogate deviation abolish keen insight lessen in intensity sour or bitter building up depart secretly renounce removal praise 3 2 3 2 2 2 2

20 COALS and the GRE

21

22

23

24

25

26

27

28

29 Task: Match each word in the first column with its definition in the second column accolade abate aberrant abscond acumen abscission acerbic accretion abjure abrogate deviation abolish keen insight lessen in intensity sour or bitter building up depart secretly renounce removal praise

30 Heuristic: competing words, pick the strongest accolade abate aberrant abscond acumen abscission acerbic accretion abjure abrogate deviation abolish keen insight lessen in intensity sour or bitter building up depart secretly renounce removal praise 7 out of 10


Download ppt "LING 581: Advanced Computational Linguistics Lecture Notes April 27th."

Similar presentations


Ads by Google