1 Gradient Grammaticality of the Indefinite Implicit Object Construction in English Tamara Nicol Medina IRCS, University of Pennsylvania Collaborators:

Slides:



Advertisements
Similar presentations
CHAPTER 2 THE NATURE OF LEARNER LANGUAGE
Advertisements

Chapter 4 Key Concepts.
Bayesian Health Technology Assessment: An Industry Statistician's Perspective John Stevens AstraZeneca R&D Charnwood Bayesian Statistics Focus Team Leader.
Defining Syntax. Lec What is Syntax? O Syntax is the scientific study of sentence structure O Science: methodology of study O Hypothesis  observation.
Movement Markonah : Honey buns, there’s something I wanted to ask you
Syntax Lecture 10: Auxiliaries. Types of auxiliary verb Modal auxiliaries belong to the category of inflection – They are in complementary distribution.
Introduction: The Chomskian Perspective on Language Study.
Learning When (and When Not) to Omit Objects in English: The Role of Verb Semantic Selectivity Tamara Nicol Medina IRCS, University of Pennsylvania Collaborators:
Introduction: The General Linear Model b b The General Linear Model is a phrase used to indicate a class of statistical models which include simple linear.
1 Language and kids Linguistics lecture #8 November 21, 2006.
Verbs and situation types continued LIN1180 Semantics Lecture 11.
Statistical Methods and Linguistics - Steven Abney Thur. POSTECH Computer Science NLP Lab Shim Jun-Hyuk.
Automatic Metaphor Interpretation as a Paraphrasing Task Ekaterina Shutova Computer Lab, University of Cambridge NAACL 2010.
Term 1 Week 9 Syntax.
9.012 Brain and Cognitive Sciences II Part VIII: Intro to Language & Psycholinguistics - Dr. Ted Gibson.
Motion Analysis (contd.) Slides are from RPI Registration Class.
Probabilistically Ranked Constraints: Derivation of the Gradient Grammaticality of Implicit Objects Tamara Nicol Medina Institute for Research in Cognitive.
Chapter Nine The Linguistic Approach: Language and Cognitive Science.
1 Simple Linear Regression Chapter Introduction In this chapter we examine the relationship among interval variables via a mathematical equation.
 2003 CSLI Publications Ling 566 Oct 16, 2007 How the Grammar Works.
1 Human simulations of vocabulary learning Présentation Interface Syntaxe-Psycholinguistique Y-Lan BOUREAU Gillette, Gleitman, Gleitman, Lederer.
Psycholinguistics 12 Language Acquisition. Three variables of language acquisition Environmental Cognitive Innate.
Lecture 1 Introduction: Linguistic Theory and Theories
1. Introduction Which rules to describe Form and Function Type versus Token 2 Discourse Grammar Appreciation.
Generative Grammar(Part ii)
TOPIC 2: Some Basic Concepts
Chapter Two Miss.Mona AL-Kahtani. Why do people study language acquisition??? Take a minute and think about it?
What makes communication by language possible? Striking fact (a) If someone utters a sentence and you know which proposition her utterance expresses, then.
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
What makes communication by language possible? Striking fact (a) If someone utters a sentence and you know which proposition her utterance expresses, then.
Albert Gatt Corpora and Statistical Methods Lecture 5.
The time-course of prediction in incremental sentence processing: Evidence from anticipatory eye movements Yuki Kamide, Gerry T.M. Altman, and Sarah L.
Emergence of Syntax. Introduction  One of the most important concerns of theoretical linguistics today represents the study of the acquisition of language.
Copyright © Cengage Learning. All rights reserved. CHAPTER 11 ANALYSIS OF ALGORITHM EFFICIENCY ANALYSIS OF ALGORITHM EFFICIENCY.
An investigation of Conservativity Tim Hunter Anastasia Conroy.
Slide 1 Definition Figures 3-4 and 3-5 Events A and B are disjoint (or mutually exclusive) if they cannot both occur together.
Tree Kernels for Parsing: (Collins & Duffy, 2001) Advanced Statistical Methods in NLP Ling 572 February 28, 2012.
Computational Lexical Semantics Lecture 8: Selectional Restrictions Linguistic Institute 2005 University of Chicago.
Using the Margins Command to Estimate and Interpret Adjusted Predictions and Marginal Effects Richard Williams
Marketing Research Aaker, Kumar, Day Seventh Edition Instructor’s Presentation Slides.
Marketing Research Aaker, Kumar, Day and Leone Tenth Edition Instructor’s Presentation Slides 1.
Trust-Aware Optimal Crowdsourcing With Budget Constraint Xiangyang Liu 1, He He 2, and John S. Baras 1 1 Institute for Systems Research and Department.
Thoughts on Model Validation for Engineering Design George A. Hazelrigg.
Statistical Estimation of Word Acquisition with Application to Readability Prediction Proceedings of the 2009 Conference on Empirical Methods in Natural.
Adele E. Goldberg. How argument structure constructions are learned.
© Child language acquisition To what extent do children acquire language by actively working out its rules?
Misuse of Articles By: Liz M. LaboyWorkshop four Albanice FloresProf. C. Garcia Jennifer M. Serrano ENGL 245.
Appraisal and Its Application to Counseling COUN 550 Saint Joseph College For Class # 3 Copyright © 2005 by R. Halstead. All rights reserved.
Models of Linguistic Choice Christopher Manning. 2 Explaining more: How do people choose to express things? What people do say has two parts: Contingent.
Cognitive Processes PSY 334 Chapter 11 – Language Structure June 2, 2003.
Linguistic Anthropology Bringing Back the Brain. What Bloomfield Got “Right” Emphasized spoken language rather than written language The role of the linguist.
Program Structure  OT Constructs formal grammars directly from markedness principles Strongly universalist: inherent typology  OT allows completely formal.
Introduction Chapter 1 Foundations of statistical natural language processing.
Chapter 8: Simple Linear Regression Yang Zhenlin.
Fita Ariyana Rombel 7 (Thursday 9 am).
1 1 Slide Simulation Professor Ahmadi. 2 2 Slide Simulation Chapter Outline n Computer Simulation n Simulation Modeling n Random Variables and Pseudo-Random.
Chapter 6 - Standardized Measurement and Assessment
 2003 CSLI Publications Ling 566 Oct 17, 2011 How the Grammar Works.
Chapter 3 Language Acquisition: A Linguistic Treatment Jang, HaYoung Biointelligence Laborotary Seoul National University.
AP Statistics From Randomness to Probability Chapter 14.
Child Syntax and Morphology
Syntax Lecture 9: Verb Types 1.
Reading and Frequency Lists
PID Controllers Jordan smallwood.
Theories of Language Development
THE NATURE of LEARNER LANGUAGE
The Nature of Learner Language (Chapter 2 Rod Ellis, 1997) Page 15
Ling 566 Oct 14, 2008 How the Grammar Works.
Traditional Grammar VS. Generative Grammar
The Nature Of Learner Language
Presentation transcript:

1 Gradient Grammaticality of the Indefinite Implicit Object Construction in English Tamara Nicol Medina IRCS, University of Pennsylvania Collaborators: Barbara Landau 1, Géraldine Legendre 1, Paul Smolensky 1, Philip Resnik 2 1 Johns Hopkins University, Department of Cognitive Science 2 University of Maryland, Department of Linguistics, Department of Computer Science

2 The (Indefinite) Implicit Object Construction (in English) John is eating John is reading Verb selects for an object, but none is overtly specified. Interpretation is of an indefinite and non- specific object. (something / some food). (something / written material). * John is reading (War and Peace). Grammaticality varies across verbs. * John is pushing. * John is opening. Verb Semantic Selectivity Aspect (Telicity, Perfectivity)

3 Overview 1. Factors that Affect Grammaticality of an Implicit Object Verb Semantic Selectivity Aspectual Properties (Telicity, Perfectivity) 2. Grammaticality Judgment Study 3. Linguistic Analysis (Optimality Theory) 4. Estimation of Constraint Ranking Probabilities 5. Implications for Acquisition

4 Verb Semantic Selectivity The omitted object tends to be recoverable from the verb. John is eating (some food) / drinking (a beverage) / singing (a song). Verbs that select for a wide variety of semantic complements, and therefore there is no one recoverable interpretation, tend to resist implicit objects. John is bringing *(something) / making *(something) / hanging *(something). Indefinite implicit objects are allowed to the extent that they are recoverable.

5 Selectional Preference Strength (SPS) (Resnik, 1996) Don’t push your brother. Move that chair. Do you want an apple? “like” Tony likes that girl. I don’t like this couch. I really like bananas. “eat” Eat your lunch. He’s eating cereal. She always eats avocados. An information-theoretic model of verbs’ strength of semantic preferences. Calculates the strength of a verb’s selection for the semantic argument classes from which its complements (or objects) are drawn. For all argument classes (c), PRIOR, Pr(c) – the overall distribution of argument classes POSTERIOR, Pr(c|v i ) – the distribution of argument classes, given a particular verb The greater the difference between Pr(c) and Pr(c|v i ), the higher SPS will be. (Argument classes were those listed in WordNet.)

6 Selectional Preference Strength (SPS) (Resnik, 1996) SPS correlated with experimental measures of recoverability and ease of inference (Resnik, 1996). –SPS corresponds to what people know about verbs’ selectional preferences. SPS correlated with rate of object omission in Brown corpus of American English (adult written English) (Resnik, 1996). –SPS directly affects syntax.

7 SPS and Implicit Objects Relative SPS is correlated with the relative frequency of an implicit object. Brown corpus of American English ( Francis and Kučera, 1982 ) SPS % Implicit Objects SPS r = 0.48, p < 0.05

8 Verb Semantic Selectivity High SPS is a necessary, but not sufficient condition on object omissibility. –Some verbs with high SPS do not occur with implicit objects, e.g., hang. –Not an inviolable rule. SPS is a continuous measure. How to incorporate this into a formal grammar? –As a statistical component to the grammar.

9 TELIC Existence of an inherent endpoint. ATELIC No inherent endpoint. “The ship sank.” Telicity (Lexical Aspect) “The ship floated.” A direct object serves to measure out the event. [+ Telic] “Kim is eating an apple.” incremental T HEME (Once the apple is gone, the event is over.) [+ Atelic] “Kim is eating.” [+Telic] “Kim arrived.” Requires an overt object. Does not require an overt object.

10 Telicity (Lexical Aspect) Atelicity is a necessary, but not sufficient condition on object omissibility. –Some atelic verbs do not occur with implicit objects, e.g., push, pull. –Not an inviolable rule.

11 Perfectivity (Grammatical Aspect) [+ Perfective] “Kim had written */? (something).” [+ Imperfective] “Kim was writing.” Requires an overt object. Does not require an overt object. PERFECTIVE Perspective of event endpoint. IMPERFECTIVE Perspective of ongoing event. have + past participle “The ship has sunk.” be + “-ing” “The ship is sinking.”

12 Perfectivity (Grammatical Aspect) Imperfectivity is a necessary, but not sufficient condition on object omissibility. –Perfectivity doesn’t render a sentence with an implicit object completely ungrammatical, while Imperfectivity doesn’t necessarily make it grammatical. Michelle had written ? (something).PERFECTIVE Michelle was hearing *(something).IMPERFECTIVE –Not an inviolable rule.

13 Putting the Puzzle Together No single factor completely distinguishes verbs that omit objects from verbs that do not. –SPS continuous measure which is related to the relative frequency of an implicit object. –Some Telic verbs do allow implicit objects, while some Atelic verbs do not. Michelle packed.TELIC Michelle wanted *(something).ATELIC –Perfectivity doesn’t render a sentence with an implicit object completely ungrammatical, while Imperfectivity doesn’t necessarily make it grammatical. Michelle had written ? (something).PERFECTIVE Michelle was hearing *(something).IMPERFECTIVE

14 Method Grammaticality Judgment Study Subjects 15 monolingual adult native speakers of English Stimuli 30 verbs, 160 sentences SPS (Resnik, 1996) Telicity Perfectivity Verb-Argument Structure Sentence Type Direct ObjectExample Sentence Two-Argument Verbs (n = 30) TargetImplicit Objects Michael had brought. Michael was bringing. ControlOvert Objects Sarah had brought a gift. Sarah was bringing a gift. One-Argument Verbs (n = 10) Filler No Objects Emma had slept. Emma was sleeping. Overt Objects Andrew had slept a blanket. Andrew was sleeping a blanket.

15 Results Grammaticality Judgment Study

16 Verb Semantic Selectivity (SPS) Grammaticality Judgment Study r = 0.66, p < 0.05

17 Telicity Grammaticality Judgment Study F = , p < 0.05

18 Perfectivity Grammaticality Judgment Study F = 3.63, p = 0.06

19 Summary of Findings Grammaticality Judgment Study Gradient across verbs. Effects of Verb Semantic Selectivity (SPS), Telicity, and Perfectivity.

20 Optimality Theory (Prince and Smolensky, 1993/2004) An Optimality Theoretic Analysis Formulate conditions as violable constraints, not inviolable rules. Take advantage of the component in OT called "CON", in which constraints are ranked with respect to one another. –It is the evaluation of the output candidates against the set of ranked constraints that determines the optimal output. –This will allow some constraints to have a greater effect than others.

21 Optimality Theory (Prince and Smolensky, 1993/2004) An Optimality Theoretic Analysis A strict ranking hierarchy (as in standard OT) will be shown to be too strong. Take insights from partial ranking approaches. Furthermore, will incorporate a statistical component to the ranking of constraints, which will allow for the derivation of GRADIENT grammaticality. However…

22 OT Framework catch (x,y) x = David, y = unspecified SPS=2.47 Telic, Perfective David had caught. David had caught something. * I NTERNAL A RGUMENT (* I NT A RG ) The output must not contain an overt internal argument (direct object). * I NT A RG  F AITHFULNESS TO A RGUMENT S TRUCTURE (F AITH A RG ) An internal argument in the input must be realized by an overt object. F AITH A RG    * I NT A RG F AITH A RG   T ELIC E NDPOINT (T ELIC E ND ) The internal argument must be overtly realized in the output, given Telic aspect. P ERFECTIVE C ODA (P ERF C ODA ) The internal argument must be overtly realized in the output, given Perfective aspect. T ELIC E ND P ERF C ODA  eat (x,y) x = David, y = unspecified SPS=3.51 Atelic, Imperfective David was eating. David was eating something.

23 Ranking of Constraints catch (x,y) x = David, y = unspecified SPS=2.47 Telic, Perfective David had caught. David had caught something. * I NT A RG  F AITH A RG    T ELIC E ND P ERF C ODA   * I NT A RG F AITH A RG  p(*I » F) p(*I » T) p(*I » P) * A RG OF H IGH SPS V ERB p(*I » F) x p(*I » T) x p(*I » P) = p( *I » {F, T, P} ) p(*I » F) = p(*I » T) = p(*I » P) = p(*I » F) x p(*I » T) x 1- [ p(*I » P) ] = p( P » *I » {F, T} ) Problems How to find perfect cut off value? Strictly ranked constraints won’t give rise to gradient grammaticality. What about SPS?What is needed is a flexible ranking of constraints. Partial Ranking: One or more constraints “floats” among other ranked constraints. Current Approach: NO ranked constraints, only a floating constraint. If * I NT A RG is highest ranked, then the implicit object is optimal. If F AITH A RG is highest ranked, then the overt object is optimal. Similar for T ELIC E ND and P ERF C ODA. Linear Function: As SPS increases, so does the relative ranking of * INT ARG. Joint Probabilities = Set of Rankings (a partial ranking of constraints) For each pairwise probability, such as p(*I » F), given a total probability of 1, there is the opposite probability, 1 - p(*I » F). Incorporating these gives rise to different partial rankings with different optimal outputs. catch (x,y) x = David, y = unspecified SPS=2.47 Telic, Imperfective

24 Total Set of Possible Partial Rankings Telic Perfective Telic Imperfective Atelic Perfective Atelic Imperfective *I » {F, T, P}implicit P » *I » {F, T}overtimplicitovertimplicit T » *I » {F, P}overt implicit {T, P} » *I » Fovert implicit F » *I » {T, P}overt {F, T} » *I » Povert {F, P} » *I » Tovert {F, T, P} » *Iovert 12.5% The various combinations of pairwise rankings can be captured by 8 partial rankings. –Give rise to OVERT or IMPLICIT object output depending on the aspectual properties of the input. 12.5%25% 50% NON-equiprobability p(*I » F) = 0.75 p(*I » T) = 0.85 p(*I » P) = %63.8%41.2%75% 35.1% 28.7% 6.2% 5.1% 11.7% 2.1% 9.6% 1.7% Probability of Implicit Object Calculate the probability of an IMPLICIT object output as the total proportion of rankings that give rise to it. –This is equivalent to the grammaticality of an implicit object output. –If equiprobable: 1/8 = 12.5%. Calculate the probability of an IMPLICIT object output as the total proportion of rankings that give rise to it. –This is equivalent to the grammaticality of an implicit object output. –If equiprobable: 1/8 = 12.5%. –But they are not equiprobable, since they depend on the joint pairwise ranking probabilities that compose them, and these are tied to SPS.

25 Summary of OT Analysis The grammaticality of an implicit object for a particular verb… is equivalent to the probability of the implicit object output for that input, which… depends upon the probabilities of each of the possible partial rankings, which… depends on the probabilities of *I » F, *I » T, and *I » P, which… are a function of SPS.

26 Finding the Probabilities So what are the pairwise probabilities of *I » F, *I » T, and *I » P in English? Can we even find probabilities that would work for all verbs? Use grammaticality judgment data to estimate the probabilities.

27 Estimation of the Constraint Rankings for English = p(*I » F)  p(*I » T)  p(*I » P) p(implicit) Telic Perfective = p(*I » {F, T, P}) = grammaticality judgment1.93 = x x.23

28 Estimated Probability Functions for English p(*I » F)p(*I » T)p(*I » P) Taking the grammaticality judgments as a direct reflection of the probabilities of an implicit object being generated by the grammar. Estimated what the pairwise rankings must be in order to produce these results. The probability of * I NT A RG ranked above each of the other three constraints increased with SPS. Steepest function for the relative ranking of * I NT A RG with T ELIC E ND.

29 Overall Predicted Grammaticality of An Implicit Object Best for Atelic Imperfective, worst for Telic Perfective. Increase as a function of SPS, but differentially depending on aspect type. -Telic Imperfectives show greatest effect of SPS.

30 Correlations between Judgments and Model Telic Perfective r = 0.84, p < 0.05 Telic Imperfective r = 0.88, p < 0.05 Atelic Imperfective r = -0.09, p > 0.05 Atelic Perfective r = 0.26, p > 0.05

31 What is the nature of the indefinite implicit object construction in the adult grammar? OT Analysis The grammaticality of an implicit object across verbs is –Gradient. –Reduced in accordance with SPS, Telicity, and Perfectivity. For any verb, if you know SPS, Telicity, and Perfectivity, then the grammar generates a relative grammaticality for the implicit object output with that verb.

32 Linguistic Analysis Turning to acquisition, we can now ask what the learner’s task must involve: Find p(*I » F), p(*I » T), and p(*I » P). How? The model’s values were estimated from grammaticality judgments. But children don’t “hear” grammaticality judgments! -Occurrence of implicit indefinite objects: increase ranking of * I NT A RG. -Occurrence of overt indefinite objects: reduce ranking of * I NT A RG.

33 Implications for Acquisition For example, Assign a grammaticality of 0 for any verb that never occurs with an implicit object. Assign a grammaticality of 1 for any verb that occurs with an implicit object at least 20% of the time. Assign a grammaticality of 0.50 for any verb that occurs with an implicit object infrequently: 0 – 20% of the time.

34 Conclusions The grammaticality of the indefinite implicit object construction is –Gradient, as shown in the Grammaticality Judgment Study. –Determined by a combination of factors, including Verb Semantic Selectivity (SPS), Telicity, and Perfectivity. It is possible to derive gradient grammaticality, by allowing constraints to "float" and assessing grammaticality over the total set of possible rankings. Estimation of the constraint ranking probabilities for English showed that it is, in fact, possible to find rankings that capture the phenomenon with low error. Raises interesting questions for acquisition: –What is the state of the child's early grammar? –How does the learner adjust her grammar in accordance with what she hears in the child-directed input (not grammaticality judgments) in order to arrive at a grammar that displays gradient judgments?