LINGUISTICA GENERALE E COMPUTAZIONALE ANALISI SINTATTICA (PARSING)

Slides:



Advertisements
Similar presentations
Natural Language Processing - Parsing 1 - Language, Syntax, Parsing Problems in Parsing Ambiguity, Attachment / Binding Bottom vs. Top Down Parsing.
Advertisements

Chapter 9: Parsing with Context-Free Grammars
PARSING WITH CONTEXT-FREE GRAMMARS
Parsing with Context Free Grammars Reading: Chap 13, Jurafsky & Martin
CKY Parsing Ling 571 Deep Processing Techniques for NLP January 12, 2011.
Artificial Intelligence 2004 Natural Language Processing - Syntax and Parsing - Language, Syntax, Parsing Problems in Parsing Ambiguity, Attachment.
1 Earley Algorithm Chapter 13.4 October 2009 Lecture #9.
 Christel Kemke /08 COMP 4060 Natural Language Processing PARSING.
CS Basic Parsing with Context-Free Grammars.
Parsing context-free grammars Context-free grammars specify structure, not process. There are many different ways to parse input in accordance with a given.
Albert Gatt LIN3022 Natural Language Processing Lecture 8.
Parsing with CFG Ling 571 Fei Xia Week 2: 10/4-10/6/05.
Context-Free Parsing. 2/37 Basic issues Top-down vs. bottom-up Handling ambiguity –Lexical ambiguity –Structural ambiguity Breadth first vs. depth first.
CMSC 723 / LING 645: Intro to Computational Linguistics November 10, 2004 Lecture 10 (Dorr): CFG’s (Finish Chapter 9) Parsing (Chapter 10) Prof. Bonnie.
Basic Parsing with Context- Free Grammars 1 Some slides adapted from Julia Hirschberg and Dan Jurafsky.
Parsing — Part II (Ambiguity, Top-down parsing, Left-recursion Removal)
CS 4705 Lecture 7 Parsing with Context-Free Grammars.
Syntactic Parsing with CFGs CMSC 723: Computational Linguistics I ― Session #7 Jimmy Lin The iSchool University of Maryland Wednesday, October 14, 2009.
CS 4705 Basic Parsing with Context-Free Grammars.
Artificial Intelligence 2004 Natural Language Processing - Syntax and Parsing - Language Syntax Parsing.
Parsing SLP Chapter 13. 7/2/2015 Speech and Language Processing - Jurafsky and Martin 2 Outline  Parsing with CFGs  Bottom-up, top-down  CKY parsing.
Syntax Construction of phrases and sentences from morphemes and words. Usually the word syntax refers to the way words are arranged together. Syntactic.
Basic Parsing with Context- Free Grammars 1 Some slides adapted from Julia Hirschberg and Dan Jurafsky.
Context-Free Grammar CSCI-GA.2590 – Lecture 3 Ralph Grishman NYU.
1 Basic Parsing with Context Free Grammars Chapter 13 September/October 2012 Lecture 6.
11 CS 388: Natural Language Processing: Syntactic Parsing Raymond J. Mooney University of Texas at Austin.
LIN LIN6932: Topics in Computational Linguistics Hana Filip.
1 Basic Parsing with Context- Free Grammars Slides adapted from Dan Jurafsky and Julia Hirschberg.
TEORIE E TECNICHE DEL RICONOSCIMENTO Linguistica computazionale in Python: -Analisi sintattica (parsing)
CS 4705 Parsing More Efficiently and Accurately. Review Top-Down vs. Bottom-Up Parsers Left-corner table provides more efficient look- ahead Left recursion.
中文信息处理 Chinese NLP Lecture 9.
1 CKY and Earley Algorithms Chapter 13 October 2012 Lecture #8.
Chapter 10. Parsing with CFGs From: Chapter 10 of An Introduction to Natural Language Processing, Computational Linguistics, and Speech Recognition, by.
10. Parsing with Context-free Grammars -Speech and Language Processing- 발표자 : 정영임 발표일 :
Fall 2004 Lecture Notes #4 EECS 595 / LING 541 / SI 661 Natural Language Processing.
Chapter 13: Parsing with Context-Free Grammars Heshaam Faili University of Tehran.
11 Syntactic Parsing. Produce the correct syntactic parse tree for a sentence.
May 2006CLINT-LN Parsing1 Computational Linguistics Introduction Parsing with Context Free Grammars.
October 2005csa3180: Parsing Algorithms 11 CSA350: NLP Algorithms Sentence Parsing I The Parsing Problem Parsing as Search Top Down/Bottom Up Parsing Strategies.
Parsing with Context Free Grammars CSC 9010 Natural Language Processing Paula Matuszek and Mary-Angela Papalaskari This slide set was adapted from: Jim.
Parsing I: Earley Parser CMSC Natural Language Processing May 1, 2003.
6/2/2016CPSC503 Winter CPSC 503 Computational Linguistics Lecture 9 Giuseppe Carenini.
October 2008CSA3180: Sentence Parsing1 CSA3180: NLP Algorithms Sentence Parsing Algorithms 2 Problems with DFTD Parser.
Sentence Parsing Parsing 3 Dynamic Programming. Jan 2009 Speech and Language Processing - Jurafsky and Martin 2 Acknowledgement  Lecture based on  Jurafsky.
Natural Language - General
Basic Parsing Algorithms: Earley Parser and Left Corner Parsing
NLP. Introduction to NLP Motivation –A lot of the work is repeated –Caching intermediate results improves the complexity Dynamic programming –Building.
November 2004csa3050: Sentence Parsing II1 CSA350: NLP Algorithms Sentence Parsing 2 Top Down Bottom-Up Left Corner BUP Implementation in Prolog.
Quick Speech Synthesis CMSC Natural Language Processing April 29, 2003.
CS 4705 Lecture 10 The Earley Algorithm. Review Top-Down vs. Bottom-Up Parsers –Both generate too many useless trees –Combine the two to avoid over-generation:
csa3050: Parsing Algorithms 11 CSA350: NLP Algorithms Parsing Algorithms 1 Top Down Bottom-Up Left Corner.
Computerlinguistik II / Sprachtechnologie Vorlesung im SS 2010 (M-GSW-10) Prof. Dr. Udo Hahn Lehrstuhl für Computerlinguistik Institut für Germanistische.
CS 4705 Lecture 7 Parsing with Context-Free Grammars.
Instructor: Nick Cercone CSEB - 1 Parsing and Context Free Grammars Parsers, Top Down, Bottom Up, Left Corner, Earley.
October 2005CSA3180: Parsing Algorithms 21 CSA3050: NLP Algorithms Parsing Algorithms 2 Problems with DFTD Parser Earley Parsing Algorithm.
November 2009HLT: Sentence Parsing1 HLT Sentence Parsing Algorithms 2 Problems with Depth First Top Down Parsing.
COMP 3438 – Part II-Lecture 5 Syntax Analysis II Dr. Zili Shao Department of Computing The Hong Kong Polytechnic Univ.
November 2004csa3050: Parsing Algorithms 11 CSA350: NLP Algorithms Parsing Algorithms 1 Top Down Bottom-Up Left Corner.
PARSING David Kauchak CS159 – Fall Admin Assignment 3 Quiz #1  High: 36  Average: 33 (92%)  Median: 33.5 (93%)
Speech and Language Processing SLP Chapter 13 Parsing.
Parsing with Context Free Grammars. Slide 1 Outline Why should you care? Parsing Top-Down Parsing Bottom-Up Parsing Bottom-Up Space (an example) Top -
CSC 594 Topics in AI – Natural Language Processing
Parsing Recommended Reading: Ch th Jurafsky & Martin 2nd edition
Basic Parsing with Context Free Grammars Chapter 13
CPSC 503 Computational Linguistics
Natural Language - General
Parsing and More Parsing
CPSC 503 Computational Linguistics
CSA2050 Introduction to Computational Linguistics
Parsing I: CFGs & the Earley Parser
Presentation transcript:

LINGUISTICA GENERALE E COMPUTAZIONALE ANALISI SINTATTICA (PARSING)

RECAP: CFG Bird et al, ch. 8.3

PARSING Parsing is the process of recognizing and assigning STRUCTURE Parsing a string with a CFG: – Finding a derivation of the string consistent with the grammar – The derivation gives us a PARSE TREE

EXAMPLE (CFR LAST WEEK)

PARSING AS SEARCH Just as in the case of non-deterministic regular expressions, the main problem with parsing is the existence of CHOICE POINTS There is a need for a SEARCH STRATEGY determining the order in which alternatives are considered

TOP-DOWN AND BOTTOM-UP SEARCH STRATEGIES The search has to be guided by the INPUT and the GRAMMAR TOP-DOWN search: the parse tree has to be rooted in the start symbol S – EXPECTATION-DRIVEN parsing BOTTOM-UP search: the parse tree must be an analysis of the input – DATA-DRIVEN parsing

AN EXAMPLE OF TOP-DOWN SEARCH (IN PARALLEL)

RECURSIVE DESCENT IN NLTK P. 303: – nltk.RecursiveDescentParser(grammar) – nltk.app.rdparser()

NON-PARALLEL SEARCH If it’s not possible to examine all alternatives in parallel, it’s necessary to make further decisions: – Which node in the current search space to expand first (breadth-first or depth-first) – Which of the applicable grammar rules to expand first – Which leaf node in a parse tree to expand next (e.g., leftmost)

TOP-DOWN, DEPTH-FIRST, LEFT-TO-RIGHT

TOP-DOWN, DEPTH-FIRST, LEFT-TO-RIGHT (II)

TOP-DOWN, DEPTH-FIRST, LEFT-TO-RIGHT (III)

TOP-DOWN, DEPTH-FIRST, LEFT-TO-RIGHT (IV)

A T-D, D-F, L-R PARSER

LEFT-RECURSION A LEFT-RECURSIVE grammar may cause a T-D, D-F, L-R parser to never return Examples of left-recursive rules: – NP  NP PP – S  S and S – But also: NP  Det Nom Det  NP’s

THE PROBLEM WITH LEFT-RECURSION

LEFT-RECURSION: POOR SOLUTIONS Rewrite the grammar to a weakly equivalent one – Problem: may not get correct parse tree Limit the depth during search – Problem: limit is arbitrary

AN EXAMPLE OF BOTTOM-UP SEARCH

SHIFT-REDUCE PARSING P. 305 – nltk.app.srparser() – ShiftReduceParser(grammar)

TOP-DOWN vs BOTTOM-UP TOP-DOWN: – Only search among grammatical answers – BUT: suggests hypotheses that may not be consistent with data – Problem: left-recursion BOTTOM-UP: – Only forms hypotheses consistent with data – BUT: may suggest hypotheses that make no sense globally

LEFT-CORNER PARSING A hybrid of top-down and bottom-up parsing Strategy: don’t consider any expansion unless the current input can serve as the LEFT- CORNER of that expansion

LC PARSING IN PYTHON

FURTHER PROBLEMS IN PARSING Ambiguity – Church and Patel (1982): the number of attachment ambiguities grows like the Catalan numbers C(2) = 2, C(3) = 5, C(4) = 14, C(5) = 132, C(6) = 469, C(7) = 1430, C(8) = 4867 Avoiding reparsing

COMMON STRUCTURAL AMBIGUITIES COORDINATION ambiguity – OLD (MEN AND WOMEN) vs (OLD MEN) AND WOMEN ATTACHMENT ambiguity: – Gerundive VP attachment ambiguity I saw the Eiffel Tower flying to Paris – PP attachment ambiguity I shot an elephant in my pajamas

PP ATTACHMENT AMBIGUITY

AMBIGUITY: SOLUTIONS Use a PROBABILISTIC GRAMMAR (not covered in this module) Use semantics

AVOID RECOMPUTING INVARIANTS Consider parsing with a top-down parser the NP: – A flight from Indianapolis to Houston on TWA With the grammar rules: – NP  Det Nominal – NP  NP PP – NP  ProperNoun

INVARIANTS AND TOP-DOWN PARSING

THE EARLEY ALGORITHM

DYNAMIC PROGRAMMING A standard T-D parser would reanalyze A FLIGHT 4 times, always in the same way A DYNAMIC PROGRAMMING algorithm uses a table (the CHART) to avoid repeating work The Earley algorithm also – Does not suffer from the left-recursion problem – Solves an exponential problem in O(n 3 )

THE CHART The Earley algorithm uses a table (the CHART) of size N+1, where N is the length of the input – Table entries sit in the `gaps’ between words Each entry in the chart is a list of – Completed constituents – In-progress constituents – Predicted constituents All three types of objects are represented in the same way as STATES

THE CHART: GRAPHICAL REPRESENTATION

STATES A state encodes two types of information: – How much of a certain rule has been encountered in the input – Which positions are covered – A  , [X,Y] DOTTED RULES – VP  V NP  – NP  Det  Nominal – S   VP

EXAMPLES

SUCCESS The parser has succeeded if entry N+1 of the chart contains the state – S   , [0,N]

THE ALGORITHM The algorithm loops through the input without backtracking, at each step performing three operations: – PREDICTOR: add predictions to the chart – COMPLETER: Move the dot to the right when looked-for constituent is found – SCANNER: read in the next input word

THE ALGORITHM: CENTRAL LOOP

EARLEY ALGORITHM: THE THREE OPERATORS

EXAMPLE, AGAIN

EXAMPLE: BOOK THAT FLIGHT

EXAMPLE: BOOK THAT FLIGHT (II)

EXAMPLE: BOOK THAT FLIGHT (III)

EXAMPLE: BOOK THAT FLIGHT (IV)

CHART PARSING IN NLTK 8.4, p. 307 ff

DEPENDENCY PARSING 8.5

READINGS Bird et al, chapter 8