Fractal Composition of Meaning: Toward a Collage Theorem for Language - PowerPoint PPT Presentation

1 / 55
About This Presentation
Title:

Fractal Composition of Meaning: Toward a Collage Theorem for Language

Description:

... millions of pixels, determine transforms for target image, and store them ... Compression algorithm computes and stores locations and transforms of tiles ... – PowerPoint PPT presentation

Number of Views:91
Avg rating:3.0/5.0
Slides: 56
Provided by: csW7
Category:

less

Transcript and Presenter's Notes

Title: Fractal Composition of Meaning: Toward a Collage Theorem for Language


1
Fractal Composition of MeaningToward a Collage
Theorem for Language
  • Simon D. Levy
  • Department of Computer Science
  • Washington and Lee University
  • Lexington, VA 24450
  • http//www.cs.wlu.edu/levy

2
Part I Self-Similarity
3
(No Transcript)
4
(No Transcript)
5
(No Transcript)
6
(No Transcript)
7
(No Transcript)
8
(No Transcript)
9
(No Transcript)
10
... But I know, too,
11
That
12
the blackbird is involved in
13
what
14
... But I know, too, That the blackbird is
involved In what Wallace Stevens
I know.
15
Part II A Little Math
16
Iterated Function Systems
  • IFS Another way to make a fractal
  • Start with an arbitrary initial image
  • Apply a set of contractive affine transforms
  • Repeat until image no longer changes
  • E.g., Sierpinski Triangle...

17
Sierpinski Triangle
I.e., three half-size copies, one in each of
three corners of 0,1 2
18
Sierpinski Triangle 0 Iterations
19
Sierpinski Triangle 1 Iteration
20
Sierpinski Triangle 2 Iterations
21
Sierpinski Triangle 3 Iterations
22
Sierpinski Triangle 4 Iterations
23
Sierpinski Triangle 5 Iterations
24
Sierpinski Triangle 6 Iterations
25
Sierpinski Triangle 7 Iterations
26
Sierpinski Triangle New Initial Image
27
Sierpinski Triangle
28
Sierpinski Triangle
29
Sierpinski Triangle
30
Sierpinski Triangle
31
Sierpinski Triangle
32
Sierpinski Triangle
33
Sierpinski Triangle
34
IFS Fractals in Nature
35
Fractal Image Compression
  • Doesn't matter what image we start with
  • All information needed to represent final
    target image is contained in transforms
  • Instead of storing millions of pixels, determine
    transforms for target image, and store them
  • How to determine transforms?

36
The Collage Theorem
  • Let
  • is the attractor or fixed
    point of
  • Collage Theorem (Barnsley 1988) Given
    arbitrary target image , transforms encoding
    are s.t.
  • Use various practical methods to find

37
Practical Fractal Image Compression
  • Most real-world images are only partially
    self-similar
  • Arbitrary images can be partitioned into
    tiles, each associated with a transform.
  • Compression algorithm computes and stores
    locations and transforms of tiles

38
Practical Fractal Image Compression
39
Practical Fractal Image Compression
40
Part III Unification
41
The Two Cultures
Semantic Relations
Discrete Symbols
Grammar Rules
Logic
Linguistics
AI
Metric Spaces
Graph Structures
Continuous Vectors
Dynamical Systems
Electrical Engineering
Images
Continuous Transforms
Chaos
42
Meanings as Vectors
You shall know a word by the company it keeps
J. R. Firth
  • Vector representation of a word encodes
    co-occurrence with other words
  • Latent Semantic Analysis (Indexing) Singular
    Value Decomposition of co-occurrence matrix on
    text 300-dimensional vectors Landauer, Dumais,
    Kintsch

43
Meanings as Vectors
  • Self-Organizing Maps Collapse high-dimensional
    descriptions (binary features or real-val
    vectors) into 2-D Kohonen
  • Simple Recurrent Networks Hidden-variable
    temporal model predicting next word based on
    current 150-D vectors Elman

44
Meanings as Vectors
Fred says the woman arrived. The woman says fred
arrived. Fred loves the woman. The woman loves
fred. The woman arrived. Fred arrived.
45
Composing Meaningful Vectors
the woman.
46
Composing Meaningful Vectors
loves the woman.
47
Composing Meaningful Vectors
Fred loves the woman.
48
Part IV Conclusions
49
Advantages of Vector Representations
  • Meaning as a gradient phenomenon (semantic
    spaces vector spaces)
  • Can represent all transforms with a single
    hidden-variable non-linear equation (grammar)
  • Gradient-descent methods as learning model
  • Principled, biologically plausible alternative
    to Words and Rules approach Chomsky, Pinker,
    Fodor

50
A Collage Theorem for Language
51
A Collage Conjecture for Language
52
A Collage Conjecture for Language
53
A Collage Hypothesis for Language
54
A Collage Hypothesis for Language
55
A Collage S.W.A.G. for Language
  • Words/meanings are co-occurrence vectors.
  • Compositions of meanings are transients to
    words.
  • Correct set of transients is one for which
    word vectors form a subset of the attractor.
Write a Comment
User Comments (0)
About PowerShow.com