Presentation is loading. Please wait.

Presentation is loading. Please wait.

Two models of accentuation that the brain might use John Goldsmith University of Chicago.

Similar presentations


Presentation on theme: "Two models of accentuation that the brain might use John Goldsmith University of Chicago."— Presentation transcript:

1 Two models of accentuation that the brain might use John Goldsmith University of Chicago

2 What’s the point? n Take a step back from linguistic analysis, and ask: what is the simplest way to perform the computations that are central and important for the data of metrical systems? n What kind of […neural…] hardware would be good at performing that kind of computation?

3 An [implicit] assumption... Linguistics is an autonomous profession, with its own autonomous methods… But it does not address an autonomous reality, nor have its autonomous truth. It shares truth with all other disciplines. And...

4 …there is no certainty that linguistic methods will get us as close to the Truth as we wish. No guarantee that linguistic methods will get us arbitrarily close to the truth. Put another way, it may be that pouring more data (of a certain sort) into contemporary linguistic methods gives us a theory that is overtrained on its data. The only escape from that is to cooperate with other research traditions.

5 Also implicit... n Linguistics and non-linguists (psychologists, neurocomputational modelers) must each take a step towards each other to find a common middle ground. n This means...

6 Non-linguists... n must realize that English is an outlier among languages... Language English

7 Linguists... n must acknowledge that much of their theory-building is motivated by their everyday convenience. (For example, they strongly prefer models whose computation requires paper or a blackboard at least, but also at most.)

8

9 Two models in neurocomputing: 1. In space: lateral inhibition Work done jointly with Gary Larson. Discrete unit modeling. [ 2. In time: neural oscillation ]

10 Dynamic computational nets ¶ Brief demonstration of the program · Some background on (some aspects of) metrical theory ¸ This network model as a minimal computational model of the solution we’re looking for. ¹ Its computation of familiar cases º Interesting properties of this network: inversion and learnability » Link to neural circuitry

11 Dynamic computational nets ¶ Brief demonstration of the program · Some background on (some aspects of) metrical theory ¸ This network model as a minimal computational model of the solution we’re looking for. ¹ Its computation of familiar cases º Interesting properties of this network: inversion and learnability » Link to neural circuitry

12 Let’s look at the program --

13

14 Dynamic computational nets ¶ Brief demonstration of the program · Some background on (some aspects of) metrical theory ¸ This network model as a minimal computational model of the solution we’re looking for. ¹ Its computation of familiar cases º Interesting properties of this network: inversion and learnability » Link to neural circuitry

15 Metrical phonology: work during 1975-1985 n Mark Liberman n Liberman and Prince n Morris Halle, J.-R. Vergnaud n Alan Prince n Bruce Hayes, especially Metrical Stress Theory (1995)

16 “Create trochaic feet, from left to right” [ x x x x x x x... Patterns of alternating stress: the simplest cases

17 “Create trochaic feet, from left to right” [ x x x x x x x... SW

18 Patterns of alternating stress: the simplest cases “Create trochaic feet, from left to right” [ x x x x x x x... SW SW

19 Patterns of alternating stress: the simplest cases “Create trochaic feet, from left to right” [ x x x x x x x... SW SWSW

20 Patterns of alternating stress: the simplest cases “Create trochaic feet, from left to right” [ x x x x x x x... SW SWSW

21 Patterns of alternating stress: the simplest cases “Create trochaic feet, from left to right” [ x x x x x x x... SW SWSW... x x x ] SW

22 SW Patterns of alternating stress: The other way... n “Create trochaic feet, from right to left” x x x x x x x x ] SW

23 Patterns of alternating stress: The other way... n “Create trochaic feet, from right to left” x x x x x x x x ] SWSW

24 Patterns of alternating stress: The other way... n “Create trochaic feet, from right to left” x x x x x x x x ] SWSWSW

25 Patterns of alternating stress: The other way... n “Create trochaic feet, from right to left” x x x x x x x x ] SWSWSWSW

26 Patterns of alternating stress: The other way... n “Create trochaic feet, from right to left” [x x x x x x x x x x x ] SWSWSWSWSW

27 This is all very convenient, but... Should be be thinking about constructing structure? Computing a result? What’s the minimal way to compute the right result?

28 Dynamic computational nets ¶ Brief demonstration of the program · Some background on (some aspects of) metrical theory ¸ This network model as a minimal computational model of the solution we’re looking for. ¹ Its computation of familiar cases º Interesting properties of this network: inversion and learnability » Link to neural circuitry

29

30 Initial activation Final activation

31

32 Beta = -.9: rightward spread of activation

33 Alpha = -.9; leftward spread of activation

34

35 Dynamic computational nets ¶ Brief demonstration of the program · Some background on (some aspects of) metrical theory ¸ This network model as a minimal computational model of the solution we’re looking for. ¹ Its computation of familiar cases º Interesting properties of this network: inversion and learnability » Link to neural circuitry

36 Examples (Hayes) Pintupi (Hansen and Hansen 1969, 1978; Australia): “syllable trochees”: odd-numbered syllables (rightward); extrametrical ultima: S s S s s S s S s S s s S s S s S s S s S s S s s

37 Weri (Boxwell and Boxwell 1966, Hayes 1980, HV 1987) n Stress the ultima, plus n Stress all odd numbered syllables, counting from the end of the word

38 Warao (Osborn 1966, HV 1987) Stress penult syllable; plus all even-numbered syllables, counting from the end of the word. (Mark last syllable as extrametrical, and run.)

39 Maranungku (Tryon 1970) Stress first syllable, and All odd-numbered syllables from the beginning of the word.

40 Garawa (Furby 1974) (or Indonesian, …) n Stress on Initial syllable; n Stress on Penult; n Stress on all even-numbered syllables, counting leftward from the end; but n “Initial dactyl effect”: no stress on the second syllable permitted.

41 Seminole/Creek (Muskogean) High tone falls on the final (ultima) or penult, depending on a parity-counting procedure that starts at the beginning of the word (“parity-counting” means counting modulo 2: 1, 0, 1, 0: like counting daisy-petals). High tone is on a count of “0”. But a heavy syllable always gets count “0”. In words with only light syllables (CV): 1 0 1 0 1 0 1 1 0 1 0 1 0 1 0

42 S ] T S T Harmonic conditioning: improves well-formedness

43 Dynamic computational nets ¶ Brief demonstration of the program · Some background on (some aspects of) metrical theory ¸ This network model as a minimal computational model of the solution we’re looking for. ¹ Its computation of familiar cases º Interesting properties of this network: inversion and learnability » Link to neural circuitry

44 = Network M Input (underlying representation) is a vector Dynamics: Output is S*: equilibrium state of (1), which by definition is: (1) Hence: Quite a surprise!

45 Inversion, again -- note the near eigenvector property Dynamics: Output is S*: equilibrium state of, which by definition is: Hence: U = S 0 M*S 0 S1S1 M*S 1 S2S2 S* = S n M*S n (I is the identity matrix)

46 Fast recoverability of underlying form This means that if you take the output S* of a network of this sort, and make the output undergo the network effect once — that’s M S* — [M’s a matrix, S a vector] and subtract that from S* — that’s (I-M) S* — you reconstruct what that network’s input state was. (This would be a highly desirable property if we had designed it in!)

47 learnability

48 Dynamic computational nets ¶ Brief demonstration of the program · Some background on (some aspects of) metrical theory ¸ This network model as a minimal computational model of the solution we’re looking for. ¹ Its computation of familiar cases º Interesting properties of this network: inversion and learnability » Link to neural circuitry

49 neural circuity

50 The challenge of language: n For the hearer: he must perceive the (intended) objects in the sensory input despite the extremely impoverished evidence of them in the signal -- a task like (but perhaps harder than) visual pattern identification; n For the speaker: she must produce and utter a signal which contains enough information to permit the hearer to perceive it as a sequence of linguistic objects.

51 Visual context: Mach bands

52 Lateral inhibition In a 1- or 2-dimensional array of neurons, neurons: n a. excite very close neighbors; n b. inhibit neighbors in a wider neighborhood; n c. do not affect cells further away Activation here... excitation Region of inhibition

53 A brief run-through on lateral inhibition... n Hartline and Ratliff 1957 in the horseshoe crab (Limulus) n Lateral inhibition leads to contrast enhancement and edge detection, under a wide range of parameter settings. n Early models used non-recurrent connections; n Later models preferred recurrent patterns of activation...

54 Recurrent lateral inhibition Recurrent models include loops of activation which retain traces of the input over longer micro-periods. (Wilson and Cowan 1972; Grossberg 1973, Amari) Recurrent inhibitory loops also leads to circuits that perform (temporal) frequency detection.

55 Recurrent lateral inhibition n …also leads to winner-take-all computations, when the weight of the lateral inhibition is great. n Most importantly for us, as noted by Wilson and Cowan 1973, lateral inhibition circuits respond characteristically to spatial frequencies.

56 n Evolution of thinking about visual cell’s receptive field from simple characteristic field (Hubel & Wiesel) to spatial frequency detector (J.P. Jones and L.A. Palmer 1987 An evaluation of the two-dimensional gabor filter model of simple receptive fields in cat striate cortex. J. Neurophysiol., 58(6):1233-1258. )

57 From U. Bochum http://www.neuroinformatik.ruhr-uni- bochum.de/ini/VDM/noframes/research/computerVision/imageProcessing/wavelets/gabor/g aborFilter.html A gabor function is a product (a convolution, really) of a sine wave and a gaussian distribution: in short, gabor cell systems implement a local spatial frequency detector.

58 Spatial frequencies n A spatial sinewave...

59 n A spatial square wave...

60 Observe how a recurrent (feedback) competitive network of lateral inhibition gives rise to a pattern of spatial waves. Initially lateral inhibition gives rise to edge detection, and classic Mach band phenomena.

61

62

63

64 end

65

66

67 Hayes’s generalizations n Culminativity: each word or phrase has a single strongest syllable bearing the main stress. T RUE IF THAT SYLLABLE IS USED TO MAP A TONE MELODY (ETL). n Rhythmic distribution: syllables bearing equal levels of stress tend to occur spaced at equal distances. n Stress hierarchies (Liberman/Prince): several levels of stress n Lack of assimilation as a natural process

68 Metrical grid xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx The height of the grid marks rhythmic prominence. Each level may represent a possible rhythmic analysis (“layer”).

69 Goldsmith-Larson (dynamic computational) model Model syllables as units with an activation level; the strength of the activation level roughly corresponds to the height of the column on the metrical grid.

70 Some generalizations about prosodic systems of the world Very crude distinction between tone and non-tone languages. It’s easier to say what a tone language is; not clear that non-tone languages form a homogeneous group. They have accent/stress...

71 Light editing of Hayes’ typology of accentual systems... “Free versus fixed stress”: when is it predictable which syllable is accented. When it is predictable, what kinds of computation are necessary to make the prediction?

72 Word-based generalizations (i.e., not sensitive to word-internal morphological structure): Rhythmic versus non-rhythmic systems In rhythmic systems, there are upper limits on how many consecutive unstressed syllables there may be. The usual limit is no more than 1. And the usual limit is no more than 1 consecutive stressed syllable.

73 I

74 Hayes’s typologies n Free vs. fixed stress (predictable or not by rule) n Rhythmic versus morphological stress –Morphological: boundary-induced versus use of morphological information to resolve competition n Bounded versus Unbounded stress (length of span of unstressed syllables)

75 Is the height of a metrical column a value of a variable? n If so, this would explain the Continuous Column Constraint: a grid is ill-formed if a grid mark on level n+1 is not matched by a grid mark on level n in the same column (an effect that shows up in several environments: in stress shift, in immobility of strong beats, main stress placement, in destressing).

76 Is constituency in metrical structure strongly motivated? #(x. ) (x. ) (x. ) (x. )... # á a á a á a á a...... (x. ) (x. ) (x. ) (x. )#... á a á a á a á a # We could think of assigning trochaic feet, either from left to right or from right to left.

77 Syllable weight Syllables divided into Heavy and Light syllables, primarily by the sum of the sonority of the post-nuclear material in the syllable. Latin stress rule: n No stress on final syllables; n stress on antepenult if penult is light; else n Stress on (heavy) penult.

78 Hayes’ parametric metrical theory n Choice of foot type: –i. size (maximum: unary/binary/ternary/ unbounded) –ii. Q-sensitivity parameter –iii. Trochaic vs. iambic (S/W, W/S) n Direction of parsing: rightward, leftward n Iterative foot assignment? n Location (create new metrical layer, new layer) n Extrametricality...

79 Extrametricality n Units (segments, syllables, feet,…) can be marked as extrametrical… n if they are peripheral (at the correct periphery)… n and enough remains after they become invisible.

80 Dynamic computational networks (Goldsmith, Larson)

81

82 Goal: to find (in some sense) the minimum computation that gets maximally close to the data at hand. What structure is required in the empirically robust cases?

83

84 = Network M Input (underlying representation) is a vector Dynamics: Output is S*: equilibrium state of (1), which by definition is: (1) Hence: Quite a surprise!

85 Learnability n Larson (1992) showed that these phonological systems were highly learnable from surface data.

86 n A spatial sinewave...

87 n A spatial square wave...

88 Observe how a recurrent (feedback) competitive network of lateral inhibition gives rise to a pattern of spatial waves. Initially lateral inhibition gives rise to edge detection, and classic Mach band phenomena.

89 Moras, Syllables, and Stress Moras and syllables (sequence of CVCVCV…)


Download ppt "Two models of accentuation that the brain might use John Goldsmith University of Chicago."

Similar presentations


Ads by Google