Presentation is loading. Please wait.

Presentation is loading. Please wait.

Starting With Complex Primitives Pays Off: Complicate Locally, Simplify Globally ARAVIND K. JOSHI Department of Computer and Information Science and Institute.

Similar presentations


Presentation on theme: "Starting With Complex Primitives Pays Off: Complicate Locally, Simplify Globally ARAVIND K. JOSHI Department of Computer and Information Science and Institute."— Presentation transcript:

1 Starting With Complex Primitives Pays Off: Complicate Locally, Simplify Globally ARAVIND K. JOSHI Department of Computer and Information Science and Institute for Research in Cognitive Science CogSci 2003 Boston, August 1 2003

2 cogsci-03: 2 Outline Summary

3 cogsci-03: 3 Introduction Formal systems to specify a grammar formalism Start with primitives (basic primitive structures or building blocks) as simple as possible and then introduce various operations for constructing more complex structures Alternatively,

4 cogsci-03: 4 Introduction: CLSG Start with complex (more complicated) primitives which directly capture some crucial linguistic properties and then introduce some general operations for composing them -- Complicate Locally, Simplify Globally (CLSG CLSG approach is characterized by localizing almost all complexity in the set of primitives, a key property

5 cogsci-03: 5 Introduction: CLSG – localization of complexity Specification of the finite set of complex primitives becomes the main task of a linguistic theory CLSG pushes all dependencies to become local, i. e., they arise initially in the primitive structures to start with

6 cogsci-03: 6 Constrained formal systems: another dimension Unconstrained formal systems -- add linguistic constraints, which become, in a sense all stipulative Alternatively, start with a constrained formal system, just adequate for describing language -- formal constraints become universal, in a sense -- other linguistic constraints become stipulative Convergence: CLSG approach leads to constrained formal systems

7 cogsci-03: 7 CLSG approach CLSG approach as led to several new insights into Syntactic description Semantic composition Language generation Statistical processing Psycholinguistic properties Discourse structure CSLG approach will be described by a particular class of grammars, TAG (LTAG), which illustrate the CSLG approach to its maximum Simple examples to communicate the interplay between formal analysis and linguistic and processing issues

8 cogsci-03: 8 Context-free Grammars The domain of locality is the one level tree -- primitive building blocks CFG, G S  NP VP VP  V NP VP  VP ADV NP  DET N DET  the N  man/car V  likes ADV  passionately S NPVP man VPADV DET N passionately likes VP NPVADV N N car DET the VP NP V

9 cogsci-03: 9 Context-free Grammars The arguments of the predicate are not in the same local domain They can be brought together in the same domain -- by introducing a rule S  NP V NP However, then the structure is lost Further the local domains of a CFG are not necessarily lexicalized Domain of Locality and Lexicalization

10 cogsci-03: 10 Towards CLSG: Lexicalization Lexical item  One or more elementary structures (trees, directed acyclic graphs), which are syntactically and semantically encapsulated. Universal combining operations Grammar  Lexicon

11 cogsci-03: 11 Lexicalized Grammars Context-free grammar (CFG) CFG, G S  NP VP VP  V NP VP  VP ADV NP  Harry NP  peanuts V  likes ADV  passionately (Non-lexical) (Lexical)S NPVP Harry VPADV V NP passionately likespeanuts

12 cogsci-03: 12 Weak Lexicalization Greibach Normal Form (GNF) CFG rules are of the form A  a B 1 B 2... B n A  a This lexicalization gives the same set of strings but not the same set of trees, i.e., the same set of structural descriptions. Hence, it is a weak lexicalization.

13 cogsci-03: 13 Strong Lexicalization Same set of strings and same set of trees or structural descriptions. Tree substitution grammars (TSG) –Increased domain of locality –Substitution as the only combining operation

14 cogsci-03: 14 :: X  X  X  Substitution

15 cogsci-03: 15 Strong Lexicalization Tree substitution grammars (TSG) CFG, G S  NP VP VP  V NP NP  Harry NP  peanuts V  likes TSG, G’  1 S NP  VP V NP  likes 22 NP Harry  3 NP peanuts

16 cogsci-03: 16 Insufficiency of TSG Formal insufficiency of TSG G: S  SS (non-lexical) S  a (lexical) CFG: TSG: G’:  1 : S SS S a 2:2: S SS S a 3:3: S a

17 cogsci-03: 17 Insufficiency of TSG TSG: G’:  1 : S SS S a 2:2: S SS S a 3:3: S a  : S S S SS SS S S a a a a a G’ can generate all strings of G but not all trees of G. CFGs cannot be lexicalized by TSG’s, i.e., by substitution only.  grows on both sides of the root

18 cogsci-03: 18  X  X* X  X X  Tree  adjoined to tree  at the node labeled X in the tree  Adjoining

19 cogsci-03: 19 With Adjoining TSG: G’:  1 : S S*S a 2:2: S S a  3 : a S G: S  SS S  a Adjoining  2 to  3 at the S node, the root node and then adjoining  1 to the S node of the derived tree we have .  : S SS SS a a a CFGs can be lexicalized by LTAGs. Adjoining is crucial for lexicalization. Adjoining arises out of lexicalization

20 cogsci-03: 20 Lexicalized LTAG Finite set of elementary trees anchored on lexical items -- extended projections of lexical anchors, -- encapsulate syntactic and semantic dependencies Elementary trees: Initial and Auxiliary Operations: Substitution and Adjoining Derivation: –Derivation Tree How elementary trees are put together. –Derived tree

21 cogsci-03: 21 agreement: person, number, gender subcategorization: sleeps: null; eats: NP; gives: NP NP; thinks: S filler-gap: who did John ask Bill to invite e word order: within and across clauses as in scrambling and clitic movement function – argument: all arguments of the lexical anchor are localized Localization of Dependencies

22 cogsci-03: 22 Localization of Dependencies word-clusters (flexible idioms): non-compositional aspect take a walk, give a cold shoulder to word co-occurrences lexical semantic aspects statistical dependencies among heads anaphoric dependencies

23 cogsci-03: 23  S NP  V likes  S NP  V likes NP  e S transitive object extraction some other trees for likes: subject extraction, topicalization, subject relative, object relative, passive, etc. VP LTAG: Examples

24 cogsci-03: 24 S NP  V likes NP  e S VP S NP  V S*  think VP  V S does S* NP  who Harry Bill     LTAG: A derivation

25 cogsci-03: 25 S NP  V likes NP  e S VP S NP  V S*  think VP  V S does S* NP  who Harry Bill     substitution adjoining who does Bill think Harry likes LTAG: A Derivation

26 cogsci-03: 26 LTAG: Derived Tree S NP S V does S NP V think VP S NP V likes e VP who Harry Bill who does Bill think Harry likes

27 cogsci-03: 27 who does Bill think Harry likes  likes  who  think  Harry  does  Bill * Compositional semantics on this derivation structure * Related to dependency diagrams substitution adjoining LTAG: Derivation Tree

28 cogsci-03: 28 S a Sb S ab S a S b S a b Nested Dependencies Architecture of elementary trees  and  determines the nature of dependencies described by the TAG grammar G:   a a a…b b b

29 cogsci-03: 29 S a S b S a S b S* Architecture of elementary trees  and  determines the kinds of dependencies that can be characterized b is one level below a and to the right of the spine Crossed dependencies   

30 cogsci-03: 30 S a S b S a S b S* S a S b S S a b a a b b Linear structure Topology of Elementary Trees: Crossed dependencies

31 cogsci-03: 31 S a S b S a S b S* S a S b S S a b a a b b (Linear) Crossed Dependencies Topology of Elementary Trees: Crossed dependencies Dependencies are nested on the tree

32 cogsci-03: 32 Examples: Nested Dependencies Center embedding of relative clauses in English (1) The rat 1 the cat 2 chased 2 ate 1 the cheese Center embedding of complement clauses in German (2) Hans 1 Peter 2 Marie 3 schwimmen 3 lassen 2 sah 1 (Hans saw Peter let/make Marie swim) Important differences between (1) and (2)

33 cogsci-03: 33 Examples: Crossed Dependencies Center embedding of complement clauses in Dutch Jan 1 Piet 2 Marie 3 zag 1 laten 2 zwemmen 3 (Jan saw Piet let/make Marie swim) It is possible to obtain a wide range of complex dependencies, i.e., complex combinations of nested and crossed dependencies. Such patterns arise in word order phenomena such as scrambling and clitic movement and also due to scope ambiguities

34 cogsci-03: 34 TAGs are more powerful than CFGs, both weakly and strongly, i.e., in terms of both -- the string sets they characterize and -- the structural descriptions they support TAGs carry over all formal properties of CFGs, modified in an appropriate way -- polynomial parsing, n 6 as compared to n 3 TAGs correspond to Embedded Pushdown Automata (EPDA) in the same way as PDAs correspond to CFGs (Vijay-Shanker, 1987) LTAG: Some Formal Properties

35 cogsci-03: 35 LTAG: Some Formal Properties An EPDA is like a PDA, however, at each move it can -- create a specified (by the move) number of stacks to the left and right of the current stack and push specified information into them -- push or pop on the current stack -- at the end of the move --stack pointer moves to the top of the rightmost stack -- if a stack becomes empty it drops out

36 cogsci-03: 36 LTAG: Some Formal Properties EPDA Input tape x x x Finite Control Current stack

37 cogsci-03: 37 LTAG: Some Formal Properties EPDA Input tape x x x Finite Control old current stack x x x newly created stacks by the move

38 cogsci-03: 38 LTAG: Some Formal Properties TAGs (more precisely, languages of TAGs) belong to the class of languages called mildly context-sensitive languages (MCSL) characterized by polynomial parsing complexity grammars for the languages in this class can characterize a limited set of patterns of nested and crossed dependencies and their combinations languages in this class have the constant growth property, i.e., sentences, if arranged in increasing order of length, grow only by a bounded amount this class properly includes CFLs

39 cogsci-03: 39 LTAG: Some Formal Properties MCSL hypothesis : Natural Languages belong to MCSL Generated very fruitful research in comparing different linguistic and formal proposals discovering provable equivalences among formalisms and constrained formal systems providing new perspectives on linguistic theories and processing issues In general, leading to a fruitful interplay of formal frameworks, substantive linguistic theories, and computational and processing paradigms

40 cogsci-03: 40 Two alternate perspectives on LTAG Supertagging Flexible composition

41 cogsci-03: 41  S NP  V likes  S NP  V likes NP  e S VP Supertagging –supertag disambiguation: Two supertags for likes Elementary trees associated with a lexical item can be regarded as super parts-of-speech (super POS or supertags) associated with that item

42 cogsci-03: 42 Supertagging –supertag disambiguation Given a corpus parsed by an LTAG grammar –we have statistics of supertags -- unigram, bigram, trigram, etc. –these statistics combine the lexical statistics as well as the statistics of the constructions in which the lexical items appear Apply statistical disambiguation techniques for standard parts-of-speech (POS) such as N (noun), V(verb), P(preposition), etc. for supertagging Joshi & Srinivas (1994), Srinivas and Joshi (1998)

43 cogsci-03: 43 Supertagging the purchase price includes two ancillary companies                                On the average a lexical item has about 15 to 20 supertags

44 cogsci-03: 44 Supertagging the purchase price includes two ancillary companies                                - Select the correct supertag for each word -- shown in blue - Correct supertag for a word means the supertag that corresponds to that word in the correct parse of the sentence

45 cogsci-03: 45 Supertagging -- performance Training corpus: 1 million words Test corpus: 47,000 words Baseline: Assign the most likely supertag: 77% Trigram supertagger: 92% Srinivas (1997) Some recent results: 93% Chen & Vijay-Shanker (2000) Improvement from 77% to 93% Comparison with standard POS: over 90% to 98%

46 cogsci-03: 46 Abstract characterization of supertagging Complex (richer) descriptions of primitives (anchors) –contrary to the standard mathematical convention Associate with each primitive all information associated with it

47 cogsci-03: 47 Abstract characterization of supertagging Making descriptions of primitives more complex –increases the local ambiguity, i.e., there are more descriptions for each primitive –however, these richer descriptions of primitives locally constrain each other –analogy to a jigsaw puzzle -- the richer the description of each primitive the better –Waltz?

48 cogsci-03: 48 Complex descriptions of primitives Making the descriptions of primitives more complex –allows statistics to be computed over these complex descriptions –these statistics are more meaningful –local statistical computations over these complex descriptions lead to robust and efficient processing –Skip?

49 cogsci-03: 49 Flexible Composition  X Split  at x X X  supertree of  at X  subtree of  at X Adjoining as Wrapping

50 cogsci-03: 50  X  X X  X X   wrapped around  i.e., the two components  and  are wrapped around   supertree of  at X  subtree of  at X Flexible Composition Adjoining as Wrapping

51 cogsci-03: 51 S V NP  likes NP(wh)  e S VP S NP  V S*S*  think VP  substitution adjoining Flexible Composition Wrapping as substitutions and adjunctions NP  - We can also view this composition as  wrapped around  - Flexible composition

52 cogsci-03: 52 S* V NP  likes NP(wh)  e S VP S NP  V S*S*  think VP  substitution adjoining Flexible Composition Wrapping as substitutions and adjunctions NP    S   and  are the two components of   attached (adjoined) to the root node S of   attached (substituted) at the foot node S of  Leads to multi-component TAG (MC-TAG)

53 cogsci-03: 53  Multi-component LTAG (MC-LTAG)     The two components are used together in one composition step. Both components attach to nodes in  an elementary tree. This preserves locality. The representation can be used for both -- predicate-argument relationships -- non-p/a information such as scope, focus, etc.

54 cogsci-03: 54 Tree-Local Multi-component LTAG (MC-LTAG) - How can the components of MC-LTAG compose preserving locality of LTAG - Tree-Local MC-LTAG -- Components of a set compose only with an elementary tree or an elementary component - Flexible composition - Tree-Local MC-LTAGs are weakly equivalent to LTAGs - However, Tree-Local MC-LTAGs provide structural descriptions not obtainable by LTAGs - Increased strong generative power

55 cogsci-03: 55 Scope ambiguities: Example   S*  NP DET NN every    S*  NP DET NN some  S NP  VPVP V hates  N student N course  ( every student hates some course)

56 cogsci-03: 56 Derivation with scope information: Example   S*  NP DET NN every    S*  NP DET NN some  S NP  VPVP V hates  N student N course  ( every student hates some course)

57 cogsci-03: 57 Derivation tree with scope information: Example  (hates)  (E)  (every)  (some)  (S)  (student)  (course) 0 0 1 2. 2 2 2 ( every student hates some course) -  and  are both adjoined at the root of  (hates) - They can be adjoined in any order, thus representing the two scope readings (underspecified representation) - The scope readings represented in the LTAG derivation itself

58 cogsci-03: 58 Tree-Local MC-LTAG and flexible semantics Applications to word order variations including scrambling, clitic movement and even scope ambiguities All word order variations up to two levels of embedding (three clauses in all) can be correctly described by tree-local MC-TAGs with flexible composition– correctly means providing appropriate structural descriptions, i.e., correct semantics -- however,

59 cogsci-03: 59 Tree-Local MC-LTAG and flexible semantics Beyond two levels of embedding not all patterns of word order variation will be correctly described Joshi, Becker, and Rambow (2002) Thus the class of tree-local MC-TAG has the property that for any grammar, G, in this class, if G works up to two levels of embedding then it fails beyond two levels for at least some patterns of word order

60 cogsci-03: 60 Tree-Local MC-LTAG and flexible semantics Beyond two levels of embedding not all patterns of word order variation will be correctly described Joshi, Becker, and Rambow (2002) Thus the class of tree-local MC-TAG has the property that for any grammar, G, in this class, if G works up to two levels of embedding then it fails beyond two levels for at least some patterns of word order Main idea

61 cogsci-03: 61 Tree-Local MC-LTAG and flexible semantics Three clauses, C1, C2, and C3, each clause can be either a single elementary tree or a multi- component tree set with two components The verb in C1 takes the verb in C2 as the argument and the verb in C2 takes the verb in C3 as the argument Flexible composition allows us to compose the three clauses in three ways

62 cogsci-03: 62 Tree-Local MC-LTAG and flexible semantics Three ways of composing C1, C2, and C3 C1 C2 C3 C1 C2 C3 C1 C2 C3 (1) (2) (3) The third mode of composition is crucial for completing the proof for two levels of embedding It is not available beyond two levels, without violating semantics!

63 cogsci-03: 63 Psycholinguistic processing issues Supertagging in psycholinguistic models Processing of crossed and nested dependencies A new twist to the competence performance distinction -- a different perspective on this distinction

64 cogsci-03: 64 Supertagging in psycholinguistic models Convergence of perspectives in the roles of -- computational linguistics and psycholinguistics Due to a shift to lexical and statistical approaches to sentence processing A particular integration by Kim, Srinivas, and Trueswell (2002) from the perspective of LTAG

65 cogsci-03: 65 Supertagging in psycholinguistic models Supertagging: Much of the computational work of linguistic analysis, -- traditionally viewed as structure building -- can be viewed as lexical disambiguation Integration of supertagging in a psycholinguistic model -- One would predict that many of the initial processing commitments of syntactic analysis are made at the lexical level in the sense of supertagging

66 cogsci-03: 66 Supertagging in psycholinguistic models Integration of a constraint-based lexicalist theory (CBL), MacDonald, Pearlmutter, and Seidenberg (1994), Trueswell and Tanenhaus (1984) lexicon represented as supertags with their distribution estimated from the supertagging experiments described earlier (Srinivas (1997))

67 cogsci-03: 67 Supertagging in psycholinguistic models Distinction between PP attachment ambiguities in (1) I saw the man in the park with a telescope (2) The secretary of the general with red hair Two supertags for with NP NP*PP NPP with VP VP*PP NPP with

68 cogsci-03: 68 Supertagging in psycholinguistic models Distinction between PP attachment ambiguities in (1) I saw the man in the park with a telescope (2) The secretary of the general with red hair In (1) the ambiguity is lexical in the supertagging sense In (2) the ambiguity is resolved at the level of attachment computation (structure building) The ambiguity in (1) is resolved at an earlier level of processing while in (2) it is resolved at a later level of processing

69 cogsci-03: 69 Supertagging in psycholinguistic models (3) The student forgot her name (4) The student forgot that the homework was due today In (1) forgot takes an NP complement while in (2) it takes a that S complement -- Thus there will be two different supertags for forgot -- The ambiguity in (3) and (4) is lexical (supertagging sense) and need not be viewed as a structural ambiguity Kim, Srinivas, and Trueswell (2002) present a neural net based architecture using supertags and confirm these and related results

70 cogsci-03: 70 Processing of nested and crossed dependencies CFG – associated automata PDA TAG – associated automata EPDA (embedded PDA) (Vijay-Shanker (1987)) EPDAs provide a new perspective on the relative ease or difficulty of processing crossed and nested dependencies which arise in -- center embedded complement constructions

71 cogsci-03: 71 Processing of nested and crossed dependencies (1)Hans 1 Peter 2 Marie 3 schwimmen 3 lassen 2 sah 1 (German– nested order) (2)Jan 1 Piet 2 Marie 3 zag 1 laten 2 zwemmen 3 (Dutch– crossed order) (3) Jan saw Peter let/make Mary swim (English– iterated order, no center embedding) Center embedding of complements -- each verb is embedded in a higher verb, except the matrix verb (top level tensed verb)

72 cogsci-03: 72 Processing of nested and crossed dependencies (1)Hans 1 Peter 2 Marie 3 schwimmen 3 lassen 2 sah 1 (German– nested order) (2)Jan 1 Piet 2 Marie 3 zag 1 laten 2 zwemmen 3 (Dutch– crossed order) Bach, Brown, and Marslen-Wilson (1986) Stated very simply, they showed that Dutch is easier than German Crossed order is easier to process than nested order

73 cogsci-03: 73 Processing of nested and crossed dependencies “ German and Dutch subjects performed two tasks -- rating comprehensibility and a test of successful comprehension—on matched sets of sentences which varied in complexity from a simple sentence to one containing three levels of embedding” “no difference between Dutch and German for sentences within the normal range (up to one level) but with a significant preference emerging for the Dutch crossed order” Bach. Brown, and Marslen-Wilson (1986)

74 cogsci-03: 74 Processing of nested and crossed dependencies (1)Hans 1 Peter 2 Marie 3 schwimmen 3 lassen 2 sah 1 (German– nested order) (2)Jan 1 Piet 2 Marie 3 zag 1 laten 2 zwemmen 3 (Dutch– crossed order) It is not enough to locate a well formed structure but we need to have a place for it to go -- In (1) a PDA can locate the innermost N 3 and V 3 but we do not know at this stage where this structure belongs, we do not have the higher verb, V 2 PDA is inadequate for (1) and, of course, for (2)

75 cogsci-03: 75 Processing of nested and crossed dependencies (1)Hans 1 Peter 2 Marie 3 schwimmen 3 lassen 2 sah 1 (German– nested order) (2)Jan 1 Piet 2 Marie 3 zag 1 laten 2 zwemmen 3 (Dutch– crossed order) EPDA can precisely model the processing of (1) and (2), consistent with the principle that -- when a well formed structure is identified it is POPPED only if there is a place for it to go, i.e., the structure in which it fits has already been POPPED --Principle of Partial Interpretation (PPI), Joshi(1990), based on Bach, Brown, and Marslen-Wilson (1986)

76 cogsci-03: 76 Processing of nested and crossed dependencies (1)Hans 1 Peter 2 Marie 3 schwimmen 3 lassen 2 sah 1 (German– nested order) (2)Jan 1 Piet 2 Marie 3 zag 1 laten 2 zwemmen 3 (Dutch– crossed order) Measure of complexity– maximum number of items from the input that have to be held back before the sentence processing (interpretation) is complete. German is about twice as hard as Dutch

77 cogsci-03: 77 Processing of nested and crossed dependencies Principle of partial interpretation (PPI) can be correctly instantiated for both Dutch and German, resulting in complexity for German about twice as that for Dutch Among all possible strategies consistent with PPI we choose the one, say, M1, which makes Dutch as hard as possible Among all possible strategies consistent with PPI we choose the one, say, M2, which makes German as easy as possible Then show that the complexity of M1 is less than M2 by about the same proportion as in Bach et al. (1986)!

78 cogsci-03: 78 Processing of nested and crossed dependencies Significance of the EPDA modeling of the processing of nested and crossed dependencies Precise correspondence between EPDA and TAG, -- direct correspondence between processing and grammars We have a precise characterization of the computational power of the processing strategy Much more recent work, e.g., Gibson (2000), Lewis (2002), Vasishth (2002)

79 cogsci-03: 79 Competence performance distinction- a new twist How do we decide whether a certain property is a competence property or a performance property? Main point: The answer depends on the formal devices available for describing language! In the context of MC-LTAG describing a variety of word order phenomena, such as scrambling, clitic movement, and even scope ambiguities, there is an interesting answer We will look at scrambling (e.g., in German)

80 cogsci-03: 80 Competence performance distinction- a new twist (1) Hans 1 Peter 2 Marie 3 schwimmen 3 lassen 2 sah 1 (Hans saw Peter make Marie swim) In (1) the three nouns are in the standard order It is possible for them to be in any order, in principle, keeping the verbs in the same order as in (1), for example, as in (2) (2) Hans 1 Marie 3 Peter 2 schwimmen 3 lassen 2 sah 1 In general, P(N 1, N 2 … N k ) V k V k-1 … V 1 where P is a permutation of k nouns

81 cogsci-03: 81 Competence performance distinction- a new twist (A) Sentences involving scrambling from more than two levels are difficult to interpret (B) Similar to the difficulty of processing more than two (perhaps even more than one) center embedding of relative clauses in English The rat the cat the dog chased bit ate the cheese (C) Since the difficulty in (B) is regarded as a performance property, we could also declare the difficulty in (A) also as performance property, but WAIT !

82 cogsci-03: 82 Competence performance distinction- a new twist We already know that the class of tree-local MC-TAG has the property that for any grammar G, in this class, if G works up to two levels of embedding then it fails beyond two levels for some patterns of word order by not being able to assign a correct structural description, i.e,, correct semantics Inability to assign correct structural descriptions is the reason for the processing difficulty!!

83 cogsci-03: 83 Competence performance distinction- a new twist So what should we conclude? The claim is not that we must conclude that the difficulty of processing sentences with scrambling from more than two levels of embedding has to be a competence property The claim is that we are presented with a choice -- the property can be a competence property -- or, we can continue to regard it a performance property

84 cogsci-03: 84 Competence performance distinction- a new twist To the best of knowledge, this is the first example where a particular processing difficulty can be claimed as a competence property Hence, whether a property is competence property or a performance property depends on the formal devices (grammars and machines) available to us for describing language What about the difficulty of processing more than two levels (perhaps only one) of center embedding of relative clauses in English?

85 cogsci-03: 85 Competence performance distinction- a new twist In order to show that the difficulty of processing sentences with more than two levels of center embeddings of relative clauses, we will have to exhibit a class of grammars, say, , such that for any grammar, G, in , if G assigns correct correct structural descriptions (correct semantics), for all sentences up to two levels of embedding, then - G fails to assign correct structural descriptions to some sentences with more than two embeddings

86 cogsci-03: 86 Competence performance distinction- a new twist For each grammar, G, in  -- if G works up to two levels then -- G fails beyond two levels However, as far as I know, we cannot exhibit such a class,   -- Finite State Grammars (FSG) will not work -- CFGs will not work -- TAGs will not work So we have no choice but to regard the processing difficulty as a performance property

87 cogsci-03: 87 Competence performance distinction- a new twist For center embedding of relative clauses -- we have no choice, so far For scrambling of center embedded complement clauses, -- we have a choice, we have an opportunity to claim the property as a competence property The two constructions are quite different The traditional assumption that all such properties have to be performance properties is not justified at all!

88 cogsci-03: 88 Summary

89 cogsci-03: 89 Tree-Local Multi-component LTAG (MC-LTAG) - How can the components of MC-LTAG compose preserving locality of LTAG - Tree-Local MC-LTAG -- Components of a set compose only with an elementary tree or an elementary component - Non-directional composition - Tree-Local MC-LTAGs are weakly equivalent to LTAGs - However, Tree-Local MC-LTAGs provide structural descriptions not obtainable by LTAGs - Increased strong generative power

90 cogsci-03: 90 Scrambling: N3 N2 N1 V3 V2 V1 VP N3 VP VP N3 e V3 VP N2 VP VP N2 V3 e VP N1 VP VP N1 V1 e VP

91 cogsci-03: 91 Scrambling: N3 N2 N1 V3 V2 V1 VP N3 VP VP N3 e V3 VP N2 VP VP N2 V2 e VP N1 VP VP N1 V1 e VP substitutionadjoining (non-directional composition, semantics of attachments)

92 cogsci-03: 92 Scrambling: N0 N3 N2 N1 V3 V2 V1 V0 VP N3 VP VP N3 e V3 VP N2 VP VP N2 V2 e VP N1 VP VP N1 V1 e VP substitutionadjoining VP N0 VP VP N0 V0 e VP ( breakdown after two levels of embedding)

93 cogsci-03: 93 Scrambling: N0 N3 N2 N1 V3 V2 V1 V0 -- Beyond two levels of embedding semantically coherent structural descriptions cannot be assigned to all scrambled strings -- the multi-component tree for V0 is forced to combine with the VP component of the V2 tree -- the V0 tree cannot be combined with the V1 tree because the composition has to be tree local -- Similar results hold for clitic ‘movement’

94 cogsci-03: 94 Semantics who does Bill think Harry likes  eats  Harry  for  fruit substitution adjoining 1 2 2.2  breakfast 2.2 eats (x, y, e) ^ Harry (x) ^ fruit (y) ^ for (e, z) ^ breakfast (z) l 1 : eats (x, y) l 2 : Harry (x) l 3 : fruit (y) l 4 : for (e, z) l 5 : breakfast(z)

95 cogsci-03: 95 Semantics Harry eats fruit for breakfast S NP  V eats VP VP* P  for PP NP  Harry fruit     breakfast NP 


Download ppt "Starting With Complex Primitives Pays Off: Complicate Locally, Simplify Globally ARAVIND K. JOSHI Department of Computer and Information Science and Institute."

Similar presentations


Ads by Google