Simple Features for Chinese Word Sense Disambiguation Hoa Trang Dang, Ching-yi Chia, Martha Palmer, Fu- Dong Chiou Computer and Information Science University.

Slides:



Advertisements
Similar presentations
Special Topics in Computer Science Advanced Topics in Information Retrieval Lecture 10: Natural Language Processing and IR. Syntax and structural disambiguation.
Advertisements

School of something FACULTY OF OTHER School of Computing FACULTY OF ENGINEERING Chunking: Shallow Parsing Eric Atwell, Language Research Group.
Using Syntax to Disambiguate Explicit Discourse Connectives in Text Source: ACL-IJCNLP 2009 Author: Emily Pitler and Ani Nenkova Reporter: Yong-Xiang Chen.
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 2 (06/01/06) Prof. Pushpak Bhattacharyya IIT Bombay Part of Speech (PoS)
How dominant is the commonest sense of a word? Adam Kilgarriff Lexicography MasterClass Univ of Brighton.
Jing-Shin Chang National Chi Nan University, IJCNLP-2013, Nagoya 2013/10/15 ACLCLP – Activities ( ) & Text Corpora.
Multilinugual PennTools that capture parses and predicate-argument structures, and their use in Applications Martha Palmer, Aravind Joshi, Mitch Marcus,
CL Research ACL Pattern Dictionary of English Prepositions (PDEP) Ken Litkowski CL Research 9208 Gue Road Damascus,
11 Chapter 20 Computational Lexical Semantics. Supervised Word-Sense Disambiguation (WSD) Methods that learn a classifier from manually sense-tagged text.
Semantic Role Labeling Abdul-Lateef Yussiff
Annotating language data Tomaž Erjavec Institut für Informationsverarbeitung Geisteswissenschaftliche Fakultät Karl-Franzens-Universität Graz Tomaž Erjavec.
LING NLP 1 Introduction to Computational Linguistics Martha Palmer April 19, 2006.
Automatic Metaphor Interpretation as a Paraphrasing Task Ekaterina Shutova Computer Lab, University of Cambridge NAACL 2010.
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
CS Word Sense Disambiguation. 2 Overview A problem for semantic attachment approaches: what happens when a given lexeme has multiple ‘meanings’?
1 Noun Homograph Disambiguation Using Local Context in Large Text Corpora Marti A. Hearst Presented by: Heng Ji Mar. 29, 2004.
Creating a Bilingual Ontology: A Corpus-Based Approach for Aligning WordNet and HowNet Marine Carpuat Grace Ngai Pascale Fung Kenneth W.Church.
Word Sense Disambiguation CMSC 723: Computational Linguistics I ― Session #11 Jimmy Lin The iSchool University of Maryland Wednesday, November 11, 2009.
CS 4705 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised –Dictionary-based.
Dept. of Computer Science & Engg. Indian Institute of Technology Kharagpur Part-of-Speech Tagging and Chunking with Maximum Entropy Model Sandipan Dandapat.
WSD using Optimized Combination of Knowledge Sources Authors: Yorick Wilks and Mark Stevenson Presenter: Marian Olteanu.
EMPOWER 2 Empirical Methods for Multilingual Processing, ‘Onoring Words, Enabling Rapid Ramp-up Martha Palmer, Aravind Joshi, Mitch Marcus, Mark Liberman,
ELN – Natural Language Processing Giuseppe Attardi
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
Some Advances in Transformation-Based Part of Speech Tagging
“How much context do you need?” An experiment about context size in Interactive Cross-language Question Answering B. Navarro, L. Moreno-Monteagudo, E.
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
Jennie Ning Zheng Linda Melchor Ferhat Omur. Contents Introduction WordNet Application – WordNet Data Structure - WordNet FrameNet Application – FrameNet.
Word Sense Disambiguation UIUC - 06/10/2004 Word Sense Disambiguation Another NLP working problem for learning with constraints… Lluís Màrquez TALP, LSI,
11 Chapter 20 Computational Lexical Semantics. Supervised Word-Sense Disambiguation (WSD) Methods that learn a classifier from manually sense-tagged text.
Paper Review by Utsav Sinha August, 2015 Part of assignment in CS 671: Natural Language Processing, IIT Kanpur.
A resource and tool for Super Sense Tagging of Italian Texts LREC 2010, Malta – 19-21/05/2010 Giuseppe Attardi* Alessandro Lenci* + Stefano Dei Rossi*
Improving Subcategorization Acquisition using Word Sense Disambiguation Anna Korhonen and Judith Preiss University of Cambridge, Computer Laboratory 15.
SYMPOSIUM ON SEMANTICS IN SYSTEMS FOR TEXT PROCESSING September 22-24, Venice, Italy Combining Knowledge-based Methods and Supervised Learning for.
CS 4705 Lecture 19 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised.
CS774. Markov Random Field : Theory and Application Lecture 19 Kyomin Jung KAIST Nov
1 Scaling Up Word Sense Disambiguation via Parallel Texts Yee Seng Chan Hwee Tou Ng Department of Computer Science National University of Singapore.
Extract Questions from Sentences. Purpose The behavior of extracting questions from sentences can be regarded as extracting semantics or knowledge from.
Lecture 21 Computational Lexical Semantics Topics Features in NLTK III Computational Lexical Semantics Semantic Web USCReadings: NLTK book Chapter 10 Text.
Natural Language Programming David Vadas The University of Sydney Supervisor: James Curran.
Grammars Grammars can get quite complex, but are essential. Syntax: the form of the text that is valid Semantics: the meaning of the form – Sometimes semantics.
Supertagging CMSC Natural Language Processing January 31, 2006.
Part-of-speech tagging
Tools for Linguistic Analysis. Overview of Linguistic Tools  Dictionaries  Linguistic Inquiry and Word Count (LIWC) Linguistic Inquiry and Word Count.
LING 6520: Comparative Topics in Linguistics (from a computational perspective) Martha Palmer Jan 15,
CSE391 – 2005 NLP 1 Events From KRR lecture. CSE391 – 2005 NLP 2 Ask Jeeves – A Q/A, IR ex. What do you call a successful movie? Tips on Being a Successful.
CIS630 1 Penn Different Sense Granularities Martha Palmer, Olga Babko-Malaya September 20, 2004.
1 Measuring the Semantic Similarity of Texts Author : Courtney Corley and Rada Mihalcea Source : ACL-2005 Reporter : Yong-Xiang Chen.
1 Fine-grained and Coarse-grained Word Sense Disambiguation Jinying Chen, Hoa Trang Dang, Martha Palmer August 22, 2003.
Using Wikipedia for Hierarchical Finer Categorization of Named Entities Aasish Pappu Language Technologies Institute Carnegie Mellon University PACLIC.
Multilinugual PennTools that capture parses and predicate-argument structures, for use in Applications Martha Palmer, Aravind Joshi, Mitch Marcus, Mark.
Word classes and part of speech tagging. Slide 1 Outline Why part of speech tagging? Word classes Tag sets and problem definition Automatic approaches.
Lexical Acquisition of Verb Direct- Object Selectional Preferences Based on the WordNet Hierarchy Emily Shen and Sushant Prakash.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
Second Language Learning From News Websites Word Sense Disambiguation using Word Embeddings.
Word Sense and Subjectivity (Coling/ACL 2006) Janyce Wiebe Rada Mihalcea University of Pittsburgh University of North Texas Acknowledgements: This slide.
Correcting Misuse of Verb Forms John Lee, Stephanie Seneff Computer Science and Artificial Intelligence Laboratory, MIT, Cambridge ACL 2008.
LING/C SC 581: Advanced Computational Linguistics Lecture Notes Feb 17 th.
English-Korean Machine Translation System
Coarse-grained Word Sense Disambiguation
SENSEVAL: Evaluating WSD Systems
CS 388: Natural Language Processing: Word Sense Disambiguation
Lecture 21 Computational Lexical Semantics
LING/C SC 581: Advanced Computational Linguistics
WordNet: A Lexical Database for English
WordNet WordNet, WSD.
Parts of Speech Mr. White English I.
English Prepositions Prof. Walid Amer.
CS224N Section 3: Corpora, etc.
CS224N Section 3: Project,Corpora
Presentation transcript:

Simple Features for Chinese Word Sense Disambiguation Hoa Trang Dang, Ching-yi Chia, Martha Palmer, Fu- Dong Chiou Computer and Information Science University of Pennsylvania {htd, chingyc, mpalmer,

Overview Maximum entropy WSD feature types English Senseval2 verbs Chinese –Penn Chinese Treebank –People’s Daily News

English Senseval2 verbs Primarily Penn Treebank WSJ corpus WordNet 1.7 sense inventory 29 verbs 15.6 senses/verb in corpus baseline (most frequent sense) 40% best system performance 60%

Local Collocational Features (English) Collocational features for w –word w –pos of w –pos of words at positions +1, -1 relative to w –words at positions -2, -1, +1, +2 relative to w

Local Syntactic Features (English) Syntactic features –whether or not the sentence is passive –whether there is a subject, direct object, indirect object, or clausal complement –the words (if any) in the positions of subject, direct object, indirect object, particle, prepositional complement (and its object)

Local Semantic Features (English) Semantic features –a Named Entity tag (PERSON, ORGANIZATION, LOCATION) for proper nouns –WordNet synsets and hypernyms for the nouns

Overall Accuracy of System (English) Feature TypeAccuracy Collocation48.3 Collocation + Syntax53.9 Collocation + Syntax + Semantics59.0 Collocation + Topic52.9 Collocation + Syntax + Topic54.2 Collocation + Syntax + Semantics + Topic60.2

Data Preparation (Chinese) Penn Chinese Treebank (100K words) CETA (Chinese-English Translation Assistance) Dictionary 28 words (multiple verb senses, possibly other pos) 3.5 senses/word in corpus Baseline (most frequent sense) 77%

Local Collocational Features (Chinese) Collocational Features: –word –pos –word-2, word-1, word+1, word+2 –pos-1, pos+1 –followsVerb

Local Syntactic Features (Chinese) Syntactic Features: –hassubj –subj –hasobj –obj-p –obj –hasinobj –Comp-VP –VPComp –Comp-IP –hasprd

Local Semantic Features (Chinese) Semantic Features (for verbs only): generated by assigning a HowNet noun category to each subject and object –subjsem –objsem

Overall Accuracy of Maximum Entropy System (CTB) Feature TypeAccuracyStd Dev Collocation (no pos) Collocation Collocation + Syntax Collocation + Syntax + Semantics Collocation + Topic Collocation + Syntax + Topic Collocation + Syntax + Semantics + Topic Baseline76.7

Data Preparation (PDN) People’s Daily News (PDN) –Five words with low accuracy and counts in CTB subsequently sense-tagged in PDN (1M words). –About 200 sentences/word from PDN. –8.2 senses/verb in corpus –Baseline (most frequent sense) 58% –Automatic segmentation, pos-tagging, parsing

Overall Accuracy of Maximum Entropy System (PDN) Feature TypeAccuracyStd Dev Collocation (no pos) Collocation Collocation + Syntax Collocation + Syntax + Semantics Collocation + Topic Collocation + Syntax + Topic Collocation + Syntax + Semantics + Topic Baseline57.6

Conclusion Types of features that are important for English and Chinese are different. –Parse information is useful for English WSD. –Lexical collocational information may be sufficient for Chinese. Chinese word sense disambiguation addressed at segmentation level