Final Assignment Demo 11 th Nov, 2012 Deepak Suyel Geetanjali Rakshit Sachin Pawar CS 626 – Sppech, NLP and the Web.

Slides:



Advertisements
Similar presentations
Three Basic Problems Compute the probability of a text: P m (W 1,N ) Compute maximum probability tag sequence: arg max T 1,N P m (T 1,N | W 1,N ) Compute.
Advertisements

School of something FACULTY OF OTHER School of Computing FACULTY OF ENGINEERING Machine Learning PoS-Taggers COMP3310 Natural Language Processing Eric.
1 CS 388: Natural Language Processing: N-Gram Language Models Raymond J. Mooney University of Texas at Austin.
Three Basic Problems 1.Compute the probability of a text (observation) language modeling – evaluate alternative texts and models P m (W 1,N ) 2.Compute.
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 2 (06/01/06) Prof. Pushpak Bhattacharyya IIT Bombay Part of Speech (PoS)
CS626 Assignments Abhirut Gupta Mandar Joshi Piyush Dungarwal (Group No. 6)
Smoothing Techniques – A Primer
Part of Speech Tagging with MaxEnt Re-ranked Hidden Markov Model Brian Highfill.
Tagging with Hidden Markov Models. Viterbi Algorithm. Forward-backward algorithm Reading: Chap 6, Jurafsky & Martin Instructor: Paul Tarau, based on Rada.
Stock Volatility Prediction using Earnings Calls Transcripts and their Summaries Naveed Ahmad Aram Zinzalian.
FSA and HMM LING 572 Fei Xia 1/5/06.
Part-of-speech Tagging cs224n Final project Spring, 2008 Tim Lai.
Ch 10 Part-of-Speech Tagging Edited from: L. Venkata Subramaniam February 28, 2002.
1 HMM (I) LING 570 Fei Xia Week 7: 11/5-11/7/07. 2 HMM Definition and properties of HMM –Two types of HMM Three basic questions in HMM.
Language Model. Major role: Language Models help a speech recognizer figure out how likely a word sequence is, independent of the acoustics. A lot of.
1 Complementarity of Lexical and Simple Syntactic Features: The SyntaLex Approach to S ENSEVAL -3 Saif Mohammad Ted Pedersen University of Toronto, Toronto.
Context-Based Metrics For Evaluating Changes to Web Pages Thesis Defense By Suvendu Kumar Dash Texas A&M University.
Albert Gatt Corpora and Statistical Methods Lecture 9.
Natural Language Processing Assignment Group Members: Soumyajit De Naveen Bansal Sanobar Nishat.
Part-of-Speech Tagging
Lemmatization Tagging LELA /20 Lemmatization Basic form of annotation involving identification of underlying lemmas (lexemes) of the words in.
1 7-Speech Recognition (Cont’d) HMM Calculating Approaches Neural Components Three Basic HMM Problems Viterbi Algorithm State Duration Modeling Training.
Graphical models for part of speech tagging
Outline POS tagging Tag wise accuracy Graph- tag wise accuracy
7-Speech Recognition Speech Recognition Concepts
Dept. of Computer Science & Engg. Indian Institute of Technology Kharagpur Part-of-Speech Tagging for Bengali with Hidden Markov Model Sandipan Dandapat,
인공지능 연구실 정 성 원 Part-of-Speech Tagging. 2 The beginning The task of labeling (or tagging) each word in a sentence with its appropriate part of speech.
A Probabilistic Graphical Model for Joint Answer Ranking in Question Answering Jeongwoo Ko, Luo Si, Eric Nyberg (SIGIR ’ 07) Speaker: Cho, Chin Wei Advisor:
CSA2050: Introduction to Computational Linguistics Part of Speech (POS) Tagging II Transformation Based Tagging Brill (1995)
Discriminative Models for Spoken Language Understanding Ye-Yi Wang, Alex Acero Microsoft Research, Redmond, Washington USA ICSLP 2006.
1 CS 552/652 Speech Recognition with Hidden Markov Models Winter 2011 Oregon Health & Science University Center for Spoken Language Understanding John-Paul.
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 3 (10/01/06) Prof. Pushpak Bhattacharyya IIT Bombay Statistical Formulation.
Part-of-Speech Tagging Foundation of Statistical NLP CHAPTER 10.
S1: Chapter 1 Mathematical Models Dr J Frost Last modified: 6 th September 2015.
Sequence Models With slides by me, Joshua Goodman, Fei Xia.
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
8.0 Search Algorithms for Speech Recognition References: of Huang, or of Becchetti, or , of Jelinek 4. “ Progress.
13-1 Chapter 13 Part-of-Speech Tagging POS Tagging + HMMs Part of Speech Tagging –What and Why? What Information is Available? Visible Markov Models.
Cluster-specific Named Entity Transliteration Fei Huang HLT/EMNLP 2005.
Tokenization & POS-Tagging
Maximum Entropy Models and Feature Engineering CSCI-GA.2590 – Lecture 6B Ralph Grishman NYU.
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
Automatic Grammar Induction and Parsing Free Text - Eric Brill Thur. POSTECH Dept. of Computer Science 심 준 혁.
Chunk Parsing II Chunking as Tagging. Chunk Parsing “Shallow parsing has become an interesting alternative to full parsing. The main goal of a shallow.
Wei Lu, Hwee Tou Ng, Wee Sun Lee National University of Singapore
Stochastic and Rule Based Tagger for Nepali Language Krishna Sapkota Shailesh Pandey Prajol Shrestha nec & MPP.
Dynamic Tuning Of Language Model Score In Speech Recognition Using A Confidence Measure Sherif Abdou, Michael Scordilis Department of Electrical and Computer.
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
Statistical Machine Translation Part II: Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
HANGMAN OPTIMIZATION Kyle Anderson, Sean Barton and Brandyn Deffinbaugh.
1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis.
CSA2050: Introduction to Computational Linguistics Part of Speech (POS) Tagging II Transformation Based Tagging Brill (1995)
Confidence Measures As a Search Guide In Speech Recognition Sherif Abdou, Michael Scordilis Department of Electrical and Computer Engineering, University.
Stochastic Methods for NLP Probabilistic Context-Free Parsers Probabilistic Lexicalized Context-Free Parsers Hidden Markov Models – Viterbi Algorithm Statistical.
Text Summarization using Lexical Chains. Summarization using Lexical Chains Summarization? What is Summarization? Advantages… Challenges…
N-Gram Model Formulas Word sequences Chain rule of probability Bigram approximation N-gram approximation.
Group – 8 Maunik Shah Hemant Adil Akanksha Patel.
Discriminative n-gram language modeling Brian Roark, Murat Saraclar, Michael Collins Presented by Patty Liu.
Probabilistic Pronunciation + N-gram Models CMSC Natural Language Processing April 15, 2003.
Dan Roth University of Illinois, Urbana-Champaign 7 Sequential Models Tutorial on Machine Learning in Natural.
Language Identification and Part-of-Speech Tagging
A Simple Approach for Author Profiling in MapReduce
Tools for Natural Language Processing Applications
An overview of decoding techniques for LVCSR
A research literature search engine with abbreviation recognition
CSCI 5832 Natural Language Processing
Neural Language Model CS246 Junghoo “John” Cho.
Assignment Demonstration
N-Gram Model Formulas Word sequences Chain rule of probability
Presentation transcript:

Final Assignment Demo 11 th Nov, 2012 Deepak Suyel Geetanjali Rakshit Sachin Pawar CS 626 – Sppech, NLP and the Web

Assignments POS Tagger – Bigram Viterbi – Trigram Viterbi – A-Star – Bigram Discriminative Viterbi Language Model (Word Prediction) – Bigram – Trigram Yago Explorer Parser Projection and NLTK 2

POS Tagger 3

Viterbi: Generative Model 4

Discriminative Bigram Model Most probable tag sequence given word sequence: 5

A-star Heuristic A : Highest transition probability – Static score which can be found directly from the learned model B : Highest lexical probability in the given sentence – Dynamic score Min_cost = -log(A)-log(B) h(n) = Min_cost * (no. of hops till goal state) 6

Comparison of different flavours of POS Taggers POS TaggerCorrectTotalAccuracy (%) Bigram Generative Viterbi Trigram Generative Viterbi A-Star Bigram Discriminative Viterbi

Language Model 8

Next word prediction : Bigram Model Using language model on raw text Using language model on POS tagged text 9

Next word prediction : Trigram Model Using language model on raw text Using language model on POS tagged text 10

Metrics: Comparing Language Models We have used “Perplexity” for comparing two language models. – Language model using only previous word – Language model using previous word as well as POS tag of previous word Perplexity is weighted average branching factor which is calculated as, 11

Results Raw text LM : – Word Prediction Accuracy: 12.97% – Perplexity : 5451 POS tagged text LM : – Word Prediction Accuracy : 13.24% – Perplexity :

Examples Raw Text - Incorrect POS tagged Text - Correct porridgy liquid is : fertiliser AJ0_porridgy NN1_liquid is : is malt dissolve into : terms NN1_malt VVB_dissolve into : into also act as : of AV0_also VVB_act as : as 13

Examples(Contd.) about english literature : and PRP_about AJ0_english literature : literature spoken english was : literature AJ0_spoken NN1_english was : was 14

Yago Explorer 15

Yago Explorer Made use of: – WikipediaCategories – WordnetCategores, and – YagoFacts. Modified Breadth First Search (BFS). 16

Algorithm Input: Entities E1, E2 Output: Paths between E1 and E2 Procedure: 1.Find WikipediaCategories for E1 and E2. If any category matches, return 2.Find WordNetCategories for E1 and E2. If any match found, return. 3.Find YagoFacts for E1 and E2. If any match found, return 4.Expand YagoFacts for E1 and E2. For each pair of entities from E1 and E2, repeat steps

Ex:1 Narendra Modi and Indian National Congress Path from E1 : Narendra_Modi--livesIn--> Gandhinagar; Gandhinagar--category--> Indian_capital_cities; Path from E2 : Indian_National_Congress--isLocatedIn--> New_Delhi; New_Delhi--category--> Indian_capital_cities; Path from E1: Narendra_Modi--isAffiliatedTo--> Bharatiya_Janata_Party; Bharatiya_Janata_Party--category--> Political_parties_in_India; Path from E2 : Indian_National_Congress--category--> Political_parties_in_India; 18

Ex:2 Mahesh Bhupathi and Mother Teresa Path from E1 : Mahesh_Bhupathi--livesIn--> Bangalore; Bangalore--category--> Metropolitan_cities_in_India; Path from E2: Mother_Teresa--diedIn--> Kolkata; Kolkata--category--> Metropolitan_cities_in_India; Path from E1 : Mahesh_Bhupathi--hasWonPrize--> Padma_Shri; Padma_Shri--category--> Civil_awards_and_decorations_of_India; Path from E2 : Mother_Teresa--hasWonPrize--> Bharat_Ratna; Bharat_Ratna--category--> Civil_awards_and_decorations_of_India; 19

Ex:3 Michelle Obama and Frederick Jelinek Path from E1 : Michelle_Obama--graduatedFrom--> Princeton_University; Princeton_University--category--> university_ ; Path from E2 : Frederick_Jelinek--graduatedFrom--> Massachusetts_Institute_of_Technology; Massachusetts_Institute_of_Technology--category--> university_ ; 20

Ex:4 Sonia Gandhi and Benito Mussolini Path from E1 : Sonia_Gandhi--isCitizenOf--> Italy ; Italy--dealsWith--> Germany ; Germany--isLocatedIn--> Europe ; Path from E2 : Benito_Mussolini--isAffiliatedTo--> National_Fascist_Party; National_Fascist_Party--isLocatedIn--> Rome; Rome--isLocatedIn--> Europe; 21

Ex5 : Narendra Modi and Mohan Bhagwat Path from E1 : – Narendra_Modi--isAffiliatedTo-- >Bharatiya_Janata_Party ; Bharatiya_Janata_Party<--isAffiliatedTo-- Hansraj_Gangaram_Ahir ; Path from E2 : – Mohan_Bhagwat--wasBornIn-->Chandrapur ; Chandrapur<--livesIn--Hansraj_Gangaram_Ahir ; 22

Parser Projection 23

Example E: Delhi is the capital of India H: dillii bhaarat kii raajdhaani hai E-parse: [ [ [Delhi] NN ] NP [ [is] VBZ [[the] ART [capital] NN ] NP [[of] P [[India] NNP ] NP ] PP ] VP ] S H-parse: [ [ [dillii] NN ] NP [ [[[bhaarat] NNP ] NP [kii] P ] PP [raaajdhaanii] NN ] NP [hai] VBZ ] VP ] S 24

Resource and Tools Parallel corpora in two languages L 1 and L 2 Parser for langauge L 1 Word translation model A statistical model of the relationship between the syntactic structures of two different languages (can be effectively learned from a bilingual corpus by an unsupervised learning technique) 25

Challenges Conflation across languages – “goes”  “ जाता है ” Phrase to phrase translation required; some phrases are opaque to translation – E.g. Phrases like “piece of cake” Noise introduced by misalignments 26

Natural LanguageTool Kit It is a platform for building Python programs to work with human language data. It provides easy-to-use interfaces to over 50 corpora and lexical resources such as WordNet It has a suite of text processing libraries for classification, tokenization, stemming, tagging, parsing, and semantic reasoning. 27

NLTK Modules Language processing task NLTK modulesFunctionality Collocation discovery nltk.collocations t-test, chi-squared, point- wise mutual information Part-of-speech tagging nltk.tag n-gram, backoff, Brill, HMM, TnT Classificationnltk.classify, nltk.cluster decision tree, maximum entropy, naive Bayes, EM, k-means Chunkingnltk.chunk regular expression, n- gram, named-entity Parsingnltk.parse chart, feature-based, unification, probabilistic, dependency 28

NLTK Modules (Contd) Language processing taskNLTK modulesFunctionality Semantic interpretationnltk.sem, nltk.inference lambda calculus, first-order logic, model checking Evaluation metricsnltk.metrics precision, recall, agreement coefficients Probability and estimationnltk.probability frequency distributions, smoothed probability distributions Applicationsnltk.app, nltk.chat graphical concordancer, parsers, WordNet browser, chatbots Linguistic fieldworknltk.toolbox manipulate data in SIL Toolbox format 29