600.465 - Intro to NLP - J. Eisner1 Phonology [These slides are missing most examples and discussion from class …]

Slides:



Advertisements
Similar presentations
Optimality Theory (OT) Prepared and presented by: Abdullah Bosaad & Liú Chàng Spring 2011.
Advertisements

Industrial Economics (Econ3400) Week 4 August 14, 2008 Room 323, Bldg 3 Semester 2, 2008 Instructor: Dr Shino Takayama.
Optimality Theory Presented by Ashour Abdulaziz, Eric Dodson, Jessica Hanson, and Teresa Li.
Intro to NLP - J. Eisner1 Finite-State Programming Some Examples.
323 Notes on Phonemic Theory in Terms of Set Theory 1. Notes on Phonemic Theory Here I will discuss phonemic theory in terms of set theory. A phoneme is.
Intro to NLP - J. Eisner1 Learning in the Limit Golds Theorem.
Intro to NLP - J. Eisner1 Syntactic Attributes Morphology, heads, gaps, etc. Note: The properties of nonterminal symbols are often called “features.”
College of Information Technology & Design
Intro to NLP - J. Eisner1 Current & Future NLP Research A Few Random Remarks.
Intro to NLP - J. Eisner1 Finite-State Methods.
CS Morphological Parsing CS Parsing Taking a surface input and analyzing its components and underlying structure Morphological parsing:
© J. Christopher Beck Lecture 7: Shifting Bottleneck.
1 Transportation Model. 2 Basic Problem The basic idea in a transportation problem is that there are sites or sources of product that need to be shipped.
Intro to NLP - J. Eisner1 Finite-State Methods.
1 Introduction to Computability Theory Lecture3: Regular Expressions Prof. Amos Israeli.
6/10/2015CPSC503 Winter CPSC 503 Computational Linguistics Lecture 3 Giuseppe Carenini.
1 Morphological analysis LING 570 Fei Xia Week 4: 10/15/07 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
Learning Bit by Bit Class 3 – Stemming and Tokenization.
The Fundamentals: Algorithms, the Integers & Matrices.
[kmpjuteynl] [fownldi]
Decision Procedures An Algorithmic Point of View
Finite-state automata 3 Morphology Day 14 LING Computational Linguistics Harry Howard Tulane University.
Intro to NLP - J. Eisner1 Part-of-Speech Tagging A Canonical Finite-State Task.
October 2006Advanced Topics in NLP1 CSA3050: NLP Algorithms Finite State Transducers for Morphological Parsing.
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2007 Lecture4 1 August 2007.
Morphological Recognition We take each sub-lexicon of each stem class and we expand each arc (e.g. the reg-noun arc) with all the morphemes that make up.
Programming Logic and Design Fourth Edition, Introductory
SVM by Sequential Minimal Optimization (SMO)
Intro to NLP - J. Eisner1 Finite-State and the Noisy Channel.
Introduction to Linguistics Ms. Suha Jawabreh Lecture 9.
Ch 7 Slide 1  Rule ordering – when there are multiple rules in the data, we have to decide if these rules interact with each other and how to order those.
Finite State Transducers for Morphological Parsing
Comprehension & Compilation in Optimality Theory Jason Eisner Jason Eisner Johns Hopkins University July 8, 2002 — ACL.
Stocs – A Stochastic CSP Solver Bella Dubrov IBM Haifa Research Lab © Copyright IBM.
Charles Kime & Thomas Kaminski © 2008 Pearson Education, Inc. (Hyperlinks are active in View Show mode) Chapter 5 – Sequential Circuits Part 2 – Sequential.
Introduction to Linguistics Ms. Suha Jawabreh Lecture # 8.
1 © 2014 B. Wilkinson Modification date: Dec Sequential Logic Circuits Previously, we described the basic building blocks of sequential circuits,
Chapter Algorithms 3.2 The Growth of Functions 3.3 Complexity of Algorithms 3.4 The Integers and Division 3.5 Primes and Greatest Common Divisors.
CSA3050: Natural Language Algorithms Finite State Devices.
CS 232: Computer Architecture II Prof. Laxmikant (Sanjay) Kale.
November 2003CSA4050: Computational Morphology IV 1 CSA405: Advanced Topics in NLP Computational Morphology IV: xfst.
 Relation: like a function, but multiple outputs ok  Regular: finite-state  Transducer: automaton w/ outputs  b  ? a  ?  aaaaa  ?  Invertible?
ALGORITHMS.
TM Design Macro Language D and SD MA/CSSE 474 Theory of Computation.
CSA4050: Advanced Topics in NLP Computational Morphology II Introduction 2 Level Morphology.
Principles Rules or Constraints
October 2004CSA3050 NLP Algorithms1 CSA3050: Natural Language Algorithms Morphological Parsing.
Intro to NLP - J. Eisner1 Phonology [These slides are missing most examples and discussion from class …]
Intro to NLP - J. Eisner1 Finite-State and the Noisy Channel.
CS : Speech, NLP and the Web/Topics in AI Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture-19: Speech: Phonetics (Using Ananthakrishnan’s presentation.
Programming Logic and Design Fourth Edition, Comprehensive Chapter 5 Making Decisions.
Lecture 1 INTRODUCTION TO ALGORITHMS Professor Uday Reddy
Optimality Theory. Linguistic theory in the 1990s... and beyond!
Introduction It had its early roots in World War II and is flourishing in business and industry with the aid of computer.
Lecture 1 Gunjeet kaur Dronacharya group of institutions.
Intro to NLP - J. Eisner1 Building Finite-State Machines.
Two Level Morphology Alexander Fraser & Liane Guillou CIS, Ludwig-Maximilians-Universität München Computational Morphology.
CIS, Ludwig-Maximilians-Universität München Computational Morphology
Part-of-Speech Tagging
Signal processing and Networking for Big Data Applications: Lecture 9 Mix Integer Programming: Benders decomposition And Branch & Bound NOTE: To change.
[These slides are missing most examples and discussion from class …]
Speech and Language Processing
CSCI 5832 Natural Language Processing
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Classification Algorithms
Finite-State and the Noisy Channel
MA/CSSE 474 Theory of Computation
Theory of Computability
Relations.
CSCI 5832 Natural Language Processing
Presentation transcript:

Intro to NLP - J. Eisner1 Phonology [These slides are missing most examples and discussion from class …]

Intro to NLP - J. Eisner2 “kats” “dawgz” “roziz” “kisiz” Pronunciation What is Phonology?  cat + -s  dog + -s  rose + -s  kiss + -s cats dogs roses kisses Spelling How do you pronounce a sequence of morphemes? Especially, how & why do you fix up the pronunciation at the seams between morphemes? why? phonology doesn’t care about the spelling (that’s just applied morphology)

Intro to NLP - J. Eisner3 “næpt” “næbd” “nadid” “natid” Pronunciation What is Phonology?  nap + -t  nab + -t  nod + -t  knot + -t napped nabbed nodded knotted Spelling Actually, these are pronounced identically: na∂ I d thanks to the English “flapping” rule (similarly: ladder/latter, bedding/betting)

Intro to NLP - J. Eisner4 “Trisyllabic Shortening” in English What is Phonology? divine  divinity futile  futility senile  senility satire  satirical decide  decision wild  wilderness serene  serenity supreme  supremity obscene  obscenity obese  *obesity (and similarly for other vowels)

Intro to NLP - J. Eisner5 What is Phonology?  A function twixt head and lip  What class of functions is allowed ?  Differs from one language to next  Often complicated, but not arbitrary  Comp Sci: How to compute, invert, learn? Morphology (head) Phonological mapping Articulation (mouth) underlying phonemes surface phones ree-ZIYN reh-zihg-NAY-shun resign resign + -ation

Intro to NLP - J. Eisner6 Successive Fixups for Phonology  Chomsky & Halle (1968)  Stepwise refinement of a single form  How to handle “resignation” example?  That is, O = f( I ) = g 3 (g 2 (g 1 ( I )))  Function composition (e.g., transducer composition) Rule 1 Rule 2 input (I) output (O) Rule 3

Intro to NLP - J. Eisner7 How to Give Orders  Directions version:  Break two eggs into a medium mixing bowl.  Remove this tab first.  On the last day of each month, come to this office and pay your rent.  Rules version:  No running in the house is allowed.  All dogs must be on a leash.  Rent must be paid by the first day of each month.  In rules version, describe what a good solution would look like, plus a search procedure for finding the best solution). Where else have we seen this? successive fixup (derivation) successive winnowing (optimization) example courtesy of K. Crosswhite

Intro to NLP - J. Eisner8 Optimality Theory for Phonology  Prince & Smolensky (1993)  Alternative to successive fixups  Successive winnowing of candidate set Gen Constraint 1 Constraint 2 Constraint 3 input... output

Intro to NLP - J. Eisner9 Optimality Theory “Tableau” constraint would prefer A, but only allowed to break tie among B,D,E  = candidate violates constraint twice (weight 2)

Intro to NLP - J. Eisner10 Optimality Theory for Phonology Gen Constraint 1 Constraint 2 Constraint 3 input (I)... output (O) adds weights to candidates adds weights to candidates best paths (several may tie) best paths (breaks some ties)

Intro to NLP - J. Eisner11 When do we prune back to best paths?  Optimality Theory: At each intermediate stage  Noisy channel: After adding up all weights... output (O)

Intro to NLP - J. Eisner12 Why does order matter?  Optimality Theory: Each machine (FSA) can choose only among outputs that previous machines liked best  Noisy channel: Each machine (FST) alters the output produced by previous machines... output (O)

Intro to NLP - J. Eisner13 Final Remark on OT Repeated best-paths only works for a single input Better to build full FST for I  O (invertible) Can do this e.g. if every constraint is binary: Assigns each candidate either 1 star (“bad”) or 0 stars (“good”) Gen Constraint 1 Constraint 2 Constraint 3 input (I)... output (O)

Intro to NLP - J. Eisner14 Optimality Theory “Tableau” all surviving candidates violate constraint 3, so we can’t eliminate any