Download presentation
Presentation is loading. Please wait.
Published byIlene Briggs Modified over 9 years ago
1
Fractal Composition of Meaning: Toward a Collage Theorem for Language Simon D. Levy Department of Computer Science Washington and Lee University Lexington, VA 24450 http://www.cs.wlu.edu/~levy
2
Part I: Self-Similarity
10
... But I know, too,
11
That
12
the blackbird is involved in
13
what what
14
... But I know, too, That the blackbird is involved In what – Wallace Stevens I know.
15
Part II: A Little Math
16
Iterated Function Systems IFS: Another way to make a fractal Start with an arbitrary initial image Apply a set of contractive affine transforms Repeat until image no longer changes E.g., Sierpinski Triangle...
17
Sierpinski Triangle I.e., three half-size copies, one in each of three corners of [0,1] 2
18
Sierpinski Triangle: 0 Iterations
19
Sierpinski Triangle: 1 Iteration
20
Sierpinski Triangle: 2 Iterations
21
Sierpinski Triangle: 3 Iterations
22
Sierpinski Triangle: 4 Iterations
23
Sierpinski Triangle: 5 Iterations
24
Sierpinski Triangle: 6 Iterations
25
Sierpinski Triangle: 7 Iterations
26
Sierpinski Triangle: New Initial Image
27
Sierpinski Triangle
34
IFS Fractals in Nature
35
Fractal Image Compression Doesn't matter what image we start with All information needed to represent final “target” image is contained in transforms Instead of storing millions of pixels, determine transforms for target image, and store them How to determine transforms?
36
The Collage Theorem Let is the attractor or “fixed point” of Collage Theorem (Barnsley 1988): Given arbitrary target image, transforms encoding are s.t. Use various practical methods to find
37
Practical Fractal Image Compression Most real-world images are only partially self- similar Arbitrary images can be partitioned into “tiles”, each associated with a transform. Compression algorithm computes and stores locations and transforms of tiles
38
Practical Fractal Image Compression
40
Part III: Unification
41
The “Two Cultures” Discrete Symbols Semantic Relations Grammar Rules Graph Structures Continuous Vectors Metric Spaces Continuous Transforms Image s Linguistics AI Logic Dynamical Systems Chaos Electrical Engineering
42
Meanings as Vectors “You shall know a word by the company it keeps” – J. R. Firth Vector representation of a word encodes co- occurrence with other words Latent Semantic Analysis (Indexing) – Singular Value Decomposition of co- occurrence matrix on text; 300-dimensional vectors [Landauer, Dumais, Kintsch]
43
Meanings as Vectors Self-Organizing Maps – Collapse high- dimensional descriptions (binary features or real-val vectors) into 2-D [Kohonen] Simple Recurrent Networks – Hidden-variable temporal model predicting next word based on current; 150-D vectors [Elman]
44
Meanings as Vectors Fred says the woman arrived. The woman says fred arrived. Fred loves the woman. The woman loves fred. The woman arrived. Fred arrived.
45
Composing Meaningful Vectors the woman.
46
Composing Meaningful Vectors loves the woman.
47
Composing Meaningful Vectors Fred loves the woman.
48
Part IV: Conclusions
49
Advantages of Vector Representations Meaning as a gradient phenomenon (semantic spaces = vector spaces) Can represent all transforms with a single hidden-variable non-linear equation (“grammar”) Gradient-descent methods as learning model Principled, biologically plausible alternative to “Words and Rules” approach [Chomsky, Pinker, Fodor]
50
A Collage Theorem for Language
51
A Collage Conjecture for Language
53
A Collage Hypothesis for Language
55
A Collage S.W.A.G. for Language Words/meanings are co-occurrence vectors. Compositions of meanings are transients to words. “Correct” set of transients is one for which word vectors form a subset of the attractor.
Similar presentations
© 2024 SlidePlayer.com Inc.
All rights reserved.