January 12, 20001 Statistical NLP: Lecture 2 Introduction to Statistical NLP.

Slides:



Advertisements
Similar presentations
CILC2011 A framework for structured knowledge extraction and representation from natural language via deep sentence analysis Stefania Costantini Niva Florio.
Advertisements

Intro to Linguistics Class # 2 Chapter 1: What is Language?
Statistical Methods and Linguistics - Steven Abney Thur. POSTECH Computer Science NLP Lab Shim Jun-Hyuk.
What is a corpus?* A corpus is defined in terms of  form  purpose The word corpus is used to describe a collection of examples of language collected.
Advanced AI - Part II Luc De Raedt University of Freiburg WS 2004/2005 Many slides taken from Helmut Schmid.
Erasmus University Rotterdam Frederik HogenboomEconometric Institute School of Economics Flavius Frasincar.
Talking about your homework News story? –What made you choose…? One of your words? –What made you choose…? (Give your vocabulary books to another student.
1/13 Parsing III Probabilistic Parsing and Conclusions.
Grammar induction by Bayesian model averaging Guy Lebanon LARG meeting May 2001 Based on Andreas Stolcke’s thesis UC Berkeley 1994.
Part of speech (POS) tagging
تمرين شماره 1 درس NLP سيلابس درس NLP در دانشگاه هاي ديگر ___________________________ راحله مکي استاد درس: دکتر عبدالله زاده پاييز 85.
Statistical Natural Language Processing Advanced AI - Part II Luc De Raedt University of Freiburg WS 2005/2006 Many slides taken from Helmut Schmid.
Latent Semantic Analysis (LSA). Introduction to LSA Learning Model Uses Singular Value Decomposition (SVD) to simulate human learning of word and passage.
Regular Expressions and Automata Chapter 2. Regular Expressions Standard notation for characterizing text sequences Used in all kinds of text processing.
Lecture 1 Introduction: Linguistic Theory and Theories
1. Introduction Which rules to describe Form and Function Type versus Token 2 Discourse Grammar Appreciation.
Corpus Linguistics What can a corpus tell us ? Levels of information range from simple word lists to catalogues of complex grammatical structures and.
1 Statistical NLP: Lecture 13 Statistical Alignment and Machine Translation.
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
현재 언어처리 기술 현황과 통 계적 접근, 사용하는 이유 3 주 강의. The Dream It’d be great if machines could –Process our (usefully) –Translate languages accurately –Help.
McEnery, T., Xiao, R. and Y.Tono Corpus-based language studies. Routledge. Unit A 2. Representativeness, balance and sampling (pp13-21)
Measuring Hint Level in Open Cloze Questions Juan Pino, Maxine Eskenazi Language Technologies Institute Carnegie Mellon University International Florida.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
1 Statistical NLP: Lecture 10 Lexical Acquisition.
Natural Language Processing Spring 2007 V. “Juggy” Jagannathan.
COMP 791A: Statistical Language Processing
Linguistics and Language
Researching language with computers Paul Thompson.
Word Sense Disambiguation (WSD)
Push Singh & Tim Chklovski. AI systems need data – lots of it! Natural language processing: Parsed & sense-tagged corpora, paraphrases, translations Commonsense.
1 Computational Linguistics Ling 200 Spring 2006.
Using Text Mining and Natural Language Processing for Health Care Claims Processing Cihan ÜNAL
Teaching language means teaching the components of language Content (also called semantics) refers to the ideas or concepts being communicated. Form refers.
Natural Language Processing (NLP) I. Introduction II. Issues in NLP III. Statistical NLP: Corpus-based Approach.
THE BIG PICTURE Basic Assumptions Linguistics is the empirical science that studies language (or linguistic behavior) Linguistics proposes theories (models)
1 Statistical NLP: Lecture 9 Word Sense Disambiguation.
1 Ch-1: Introduction (1.3 & 1.4 & 1.5) Prepared by Qaiser Abbas ( )
CS 4705 Lecture 19 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised.
인공지능 연구실 황명진 FSNLP Introduction. 2 The beginning Linguistic science 의 4 부분 –Cognitive side of how human acquire, produce, and understand.
Natural Language Processing Spring 2007 V. “Juggy” Jagannathan.
How Solvable Is Intelligence? A brief introduction to AI Dr. Richard Fox Department of Computer Science Northern Kentucky University.
1 CSI 5180: Topics in AI: Natural Language Processing, A Statistical Approach Instructor: Nathalie Japkowicz Objectives of.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
1 Statistical NLP: Lecture 7 Collocations. 2 Introduction 4 Collocations are characterized by limited compositionality. 4 Large overlap between the concepts.
For Monday Read chapter 24, sections 1-3 Homework: –Chapter 23, exercise 8.
For Monday Read chapter 26 Last Homework –Chapter 23, exercise 7.
Introduction Chapter 1 Foundations of statistical natural language processing.
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
The Unreasonable Effectiveness of Data
Levels of Linguistic Analysis
SIMS 296a-4 Text Data Mining Marti Hearst UC Berkeley SIMS.
Corpus Linguistics MOHAMMAD ALIPOUR ISLAMIC AZAD UNIVERSITY, AHVAZ BRANCH.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
A Simple English-to-Punjabi Translation System By : Shailendra Singh.
Chapter 11 Language. Some Questions to Consider How do we understand individual words, and how are words combined to create sentences? How can we understand.
INTRODUCTION TO APPLIED LINGUISTICS
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Bayes Rule Mutual Information Conditional.
NLP Midterm Solution #1 bilingual corpora –parallel corpus (document-aligned, sentence-aligned, word-aligned) (4) –comparable corpus (4) Source.
An –Najah National University Submitted to : Dr. Suzan Arafat
Introduction to Corpus Linguistics
Statistical NLP: Lecture 7
현재 언어처리 기술 현황과 통계적 접근, 사용하는 이유
Statistical NLP: Lecture 13
Statistical NLP: Lecture 9
CS621/CS449 Artificial Intelligence Lecture Notes
Levels of Linguistic Analysis
Introduction to Artificial Intelligence Instructor: Dr. Eduardo Urbina
Information Retrieval
Statistical NLP : Lecture 9 Word Sense Disambiguation
Statistical NLP: Lecture 10
Presentation transcript:

January 12, Statistical NLP: Lecture 2 Introduction to Statistical NLP

January 12, Rational versus Empiricist Approaches to Language I 4 Question: What prior knowledge should be built into our models of NLP? 4 Rationalist Answer: A significant part of the knowledge in the human mind is not derived by the senses but is fixed in advance, presumably by genetic inheritance (Chomsky: poverty of the stimulus). 4 Empiricist Answer: The brain is able to perform association, pattern recognition, and generalization and, thus, the structures of Natural Language can be learned.

January 12, Rational versus Empiricist Approaches to Language II 4 Chomskyan/generative linguists seek to describe the language module of the human mind (the I- language) for which data such as text (the E- language) provide only indirect evidence, which can be supplemented by native speakers intuitions. 4 Empiricists approaches are interested in describing the E-language as it actually occurs. 4 Chomskyans make a distinction between linguistic competence and linguistic performance. They believe that linguistic competence can be described in isolation while Empiricists reject this notion.

January 12, Today’s Approach to NLP 4 From ~ , people were concerned with the science of the mind and built small (toy) systems that attempted to behave intelligently. 4 Recently, there has been more interest on engineering practical solutions using automatic learning (knowledge induction). 4 While Chomskyans tend to concentrate on categorical judgements about very rare types of sentences, statistical NLP practitioners concentrate on common types of sentences.

January 12, Why is NLP Difficult? 4 NLP is difficult because Natural Language is highly ambiguous. 4 Example: “Our company is training workers” has 3 parses (i.e., syntactic analyses). 4 “List the sales of the products produced in 1973 with the products produced in 1972” has 455 parses. 4 Therefore, a practical NLP system must be good at making disambiguation decisions of word sense, word category, syntactic structure, and semantic scope.

January 12, Methods that don’t work well 4 Maximizing coverage while minimizing ambiguity is inconsistent with symbolic NLP. 4 Furthermore, hand-coding syntactic constraints and preference rules are time consuming to build, do not scale up well and are brittle in the face of the extensive use of metaphor in language. 4 Example: if we code animate being --> swallow --> physical object I swallowed his story, hook, line, and sinker The supernova swallowed the planet.

January 12, What Statistical NLP can do for us 4 Disambiguation strategies that rely on hand-coding produce a knowledge acquisition bottleneck and perform poorly on naturally occurring text. 4 A Statistical NLP approach seeks to solve these problems by automatically learning lexical and structural preferences from corpora. In particular, Statistical NLP recognizes that there is a lot of information in the relationships between words. 4 The use of statistics offers a good solution to the ambiguity problem: statistical models are robust, generalize well, and behave gracefully in the presence of errors and new data.

January 12, Things that can be done with Text Corpora I: Word Counts 4 Word Counts to find out: –What are the most common words in the text. –How many words are in the text (word tokens and word types). –What the average frequency of each word in the text is. 4 Limitation of word counts: Most words appear very infrequently and it is hard to predict much about the behavior of words that do not occur often in a corpus. ==> Zipf’s Law.

January 12, Things that can be done with Text Corpora II: Zipf’s Law 4 If we count up how often each word type of a language occurs in a large corpus and then list the words in order of their frequency of occurrence, we can explore the relationship between the frequency of a word, f, and its position in the list, known as its rank, r. 4 Zipf’s Law says that: f  1/r 4 Significance of Zipf’s Law: For most words, our data about their use will be exceedingly sparse. Only for a few words will we have a lot of examples.

January 12, Things that can be done with Text Corpora III: Collocations 4 A collocation is any turn of phrase or accepted usage where somehow the whole is perceived as having an existence beyond the sum of its parts (e.g., disk drive, make up, bacon and eggs). 4 Collocations are important for machine translation. 4 Collocation can be extracted from a text (example, the most common bigrams can be extracted). However, since these bigrams are often insignificant (e.g., “at the”, “of a”), they can be filtered.

January 12, Things that can be done with Text Corpora IV: Concordances 4 Finding concordances corresponds to finding the different contexts in which a given word occurs. 4 One can use a Key Word In Context (KWIC) concordancing program. 4 Concordances are useful both for building dictionaries for learners of foreign languages and for guiding statistical parsers.