CS460/449 : Speech, Natural Language Processing and the Web/Topics in AI Programming (Lecture 2– Introduction+ML and NLP) Pushpak Bhattacharyya CSE Dept.,

Slides:



Advertisements
Similar presentations
School of something FACULTY OF OTHER School of Computing FACULTY OF ENGINEERING Chunking: Shallow Parsing Eric Atwell, Language Research Group.
Advertisements

CS344: Introduction to Artificial Intelligence
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 2 (06/01/06) Prof. Pushpak Bhattacharyya IIT Bombay Part of Speech (PoS)
CPSC 422, Lecture 16Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 16 Feb, 11, 2015.
Sequence Classification: Chunking Shallow Processing Techniques for NLP Ling570 November 28, 2011.
Semantic Role Labeling Abdul-Lateef Yussiff
CS621: Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 29– AI and Probability (exemplified through NLP) 4 th Oct, 2010.
NLP and Speech Course Review. Morphological Analyzer Lexicon Part-of-Speech (POS) Tagging Grammar Rules Parser thethe – determiner Det NP → Det.
Introduction to Computational Linguistics Lecture 2.
Ch 10 Part-of-Speech Tagging Edited from: L. Venkata Subramaniam February 28, 2002.
1 CSC 594 Topics in AI – Applied Natural Language Processing Fall 2009/ Shallow Parsing.
Introduction to CL Session 1: 7/08/2011. What is computational linguistics? Processing natural language text by computers  for practical applications.
Machine Translation Prof. Alexandros Potamianos Dept. of Electrical & Computer Engineering Technical University of Crete, Greece May 2003.
Part of speech (POS) tagging
Machine Learning in Natural Language Processing Noriko Tomuro November 16, 2006.
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2005 Lecture 1 21 July 2005.
Methods in Computational Linguistics II Queens College Lecture 5: List Comprehensions.
ELN – Natural Language Processing Giuseppe Attardi
9/8/20151 Natural Language Processing Lecture Notes 1.
Pushpak Bhattacharyya CSE Dept., IIT Bombay
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
CS344: Introduction to Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 26: Probabilistic Parsing.
Ling 570 Day 17: Named Entity Recognition Chunking.
Prof. Pushpak Bhattacharyya, IIT Bombay.1 Application of Noisy Channel, Channel Entropy CS 621 Artificial Intelligence Lecture /09/05.
10/12/2015CPSC503 Winter CPSC 503 Computational Linguistics Lecture 10 Giuseppe Carenini.
1 Statistical NLP: Lecture 9 Word Sense Disambiguation.
CSA2050: Introduction to Computational Linguistics Part of Speech (POS) Tagging II Transformation Based Tagging Brill (1995)
1 Semi-Supervised Approaches for Learning to Parse Natural Languages Rebecca Hwa
Methods for the Automatic Construction of Topic Maps Eric Freese, Senior Consultant ISOGEN International.
CS460/449 : Speech, Natural Language Processing and the Web/Topics in AI Programming (Lecture 1 – Introduction) Pushpak Bhattacharyya CSE Dept., IIT Bombay.
Introduction to CL & NLP CMSC April 1, 2003.
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 3 (10/01/06) Prof. Pushpak Bhattacharyya IIT Bombay Statistical Formulation.
CS 4705 Lecture 19 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised.
CS626: NLP, Speech and the Web Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 15, 17: Parsing Ambiguity, Probabilistic Parsing, sample seminar 17.
CS774. Markov Random Field : Theory and Application Lecture 19 Kyomin Jung KAIST Nov
A Cascaded Finite-State Parser for German Michael Schiehlen Institut für Maschinelle Sprachverarbeitung Universität Stuttgart
10/30/2015CPSC503 Winter CPSC 503 Computational Linguistics Lecture 7 Giuseppe Carenini.
1 CSI 5180: Topics in AI: Natural Language Processing, A Statistical Approach Instructor: Nathalie Japkowicz Objectives of.
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture–39: Recap.
CS : Speech, NLP and the Web/Topics in AI Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture-17: Probabilistic parsing; inside- outside probabilities.
CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 29– CYK; Inside Probability; Parse Tree construction) Pushpak Bhattacharyya CSE.
CS : Speech, NLP and the Web/Topics in AI Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture-16: Probabilistic parsing; computing probability of.
For Monday Read chapter 26 Last Homework –Chapter 23, exercise 7.
CSA3202 Human Language Technology HMMs for POS Tagging.
CS : Speech, NLP and the Web/Topics in AI Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture-14: Probabilistic parsing; sequence labeling, PCFG.
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 1 (03/01/06) Prof. Pushpak Bhattacharyya IIT Bombay Introduction to Natural.
CPSC 422, Lecture 27Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27 Nov, 16, 2015.
Shallow Parsing for South Asian Languages -Himanshu Agrawal.
CS : Speech, NLP and the Web/Topics in AI Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture-15: Probabilistic parsing; PCFG (contd.)
For Monday Read chapter 26 Homework: –Chapter 23, exercises 8 and 9.
Part-of-Speech Tagging & Sequence Labeling Hongning Wang
Overview of Statistical NLP IR Group Meeting March 7, 2006.
LING/C SC 581: Advanced Computational Linguistics Lecture Notes Feb 3 rd.
CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 25– Probabilistic Parsing) Pushpak Bhattacharyya CSE Dept., IIT Bombay 14 th March,
Dan Roth University of Illinois, Urbana-Champaign 7 Sequential Models Tutorial on Machine Learning in Natural.
Pushpak Bhattacharyya CSE Dept., IIT Bombay
Natural Language Processing (NLP)
CS : Speech, NLP and the Web/Topics in AI
CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 23, 24– Parsing Algorithms; Parsing in case of Ambiguity; Probabilistic Parsing)
Machine Learning in Natural Language Processing
CS621/CS449 Artificial Intelligence Lecture Notes
Automatic Detection of Causal Relations for Question Answering
Chunk Parsing CS1573: AI Application Development, Spring 2003
Pushpak Bhattacharyya CSE Dept., IIT Bombay
Natural Language Processing (NLP)
Artificial Intelligence 2004 Speech & Natural Language Processing
Natural Language Processing (NLP)
Presentation transcript:

CS460/449 : Speech, Natural Language Processing and the Web/Topics in AI Programming (Lecture 2– Introduction+ML and NLP) Pushpak Bhattacharyya CSE Dept., IIT Bombay

Persons involved Faculty instructor: Dr. Pushpak Bhattacharyya ( Areas of Expertise: Natural Language Processing, Machine Learning TAs: Prithviraj and Debraj Course home page (to be created) Mirrored at

Time and Venue Slot-3 Old CSE: S9 (top floor) Mo , Tu , Th- 8.30

Perpectivising NLP: Areas of AI and their inter-dependencies Search Vision Planning Machine Learning Knowledge Representation Logic Expert Systems RoboticsNLP AI is the forcing function for Computer Science

Stages of language processing Phonetics and phonology Morphology Lexical Analysis Syntactic Analysis Semantic Analysis Pragmatics Discourse

Two Views of NLP 1. Classical View: Layered Procssing;Various Ambiguities (already discussed) 2. Statistical/Machine Learning View

Uncertainty in classification: Ambiguity Visiting aunts can be a nuisance Visiting: adjective or gerund (POS tag ambiguity) Role of aunt: agent of visit (aunts are visitors) object of visit (aunts are being visited) Minimize uncertainty of classification with cues from the sentence

What cues? Position with respect to the verb: France to the left of beat and Brazil to the right: agent-object role marking (English) Case marking: France ne (Hindi); ne (Marathi): agent role Brazil ko (Hindi); laa (Marathi): object role Morphology: haraayaa (hindi); haravlaa (Marathi): verb POS tag as indicated by the distinctive suffixes

Cues are like attribute-value pairs prompting machine learning from NL data Constituent ML tasks Goal: classification or clustering Features/attributes (word position, morphology, word label etc.) Values of features Training data (corpus: annotated or un-annotated) Test data (test corpus) Accuracy of decision (precision, recall, F-value, MAP etc.) Test of significance (sample space to generality)

What is the output of an ML-NLP System (1/2) Option 1: A set of rules, e.g., If the word to the left of the verb is a noun and has animacy feature, then it is the likely agent of the action denoted by the verb. The child broke the toy (child is the agent) The window broke (window is not the agent; inanimate)

What is the output of an ML-NLP System (2/2) Option 2: a set of probability values P(agent|word is to the left of verb and has animacy) > P(object|word is to the left of verb and has animacy)> P(instrument|word is to the left of verb and has animacy) etc.

How is this different from classical NLP corpus Text data Linguist Computer rules rules/probabilities Classical NLP Statistical NLP

Classification appears as sequence labeling

A set of Sequence Labeling Tasks: smaller to larger units Words: Part of Speech tagging Named Entity tagging Sense marking Phrases: Chunking Sentences: Parsing Paragraphs: Co-reference annotating

Example of word labeling: POS Tagging Come September, and the UJF campus is abuzz with new and returning students. Come_VB September_NNP,_, and_CC the_DT UJF_NNP campus_NN is_VBZ abuzz_JJ with_IN new_JJ and_CC returning_VBG students_NNS._.

Example of word labeling: Named Entity Tagging September UJF

Example of word labeling: Sense Marking WordSynsetWN-synset-no come{arrive, get, come} abuzz{abuzz, buzzing, droning}

Example of phrase labeling: Chunking Come July, and is abuzz with. the UJF campus new and returning students

Example of Sentence labeling: Parsing [ S1 [ S [ S [ VP [ VB Come][ NP [ NNP July]]]] [,,] [ CC and] [ S [ NP [ DT the] [ JJ UJF] [ NN campus]] [ VP [ AUX is] [ ADJP [ JJ abuzz] [ PP [ IN with] [ NP [ ADJP [ JJ new] [ CC and] [ VBG returning]] [ NNS students]]]]]] [..]]]

Handling labeling through the Noisy Channel Model w t (w n, w n-1, …, w 1 ) (t m, t m-1, …, t 1 ) Noisy Channel Sequence w is transformed into sequence t.

Bayesian Decision Theory and Noisy Channel Model are close to each other Bayes Theorem : Given the random variables A and B, Posterior probability Prior probability Likelihood

Corpus A collection of text called corpus, is used for collecting various language data With annotation: more information, but manual labor intensive Practice: label automatically; correct manually The famous Brown Corpus contains 1 million tagged words. Switchboard: very famous corpora 2400 conversations, 543 speakers, many US dialects, annotated with orthography and phonetics

Example-1 of Application of Noisy Channel Model: Probabilistic Speech Recognition (Isolated Word)[8] Problem Definition : Given a sequence of speech signals, identify the words. 2 steps : Segmentation (Word Boundary Detection) Identify the word Isolated Word Recognition : Identify W given SS (speech signal)

Identifying the word P(SS|W) = likelihood called “phonological model “  intuitively more tractable! P(W) = prior probability called “language model”

Pronunciation Dictionary P(SS|W) is maintained in this way. P(t o m ae t o |Word is “tomato”) = Product of arc probabilities t s4s4 omo ae t aa end s1s1 s2s2 s3s3 s5s5 s6s6 s7s Word Pronunciation Automaton Tomato

Discriminative vs. Generative Model W * = argmax (P(W|SS)) W Compute directly from P(W|SS) Compute from P(W).P(SS|W) Discriminative Model Generative Model