Kees van Deemter - PowerPoint PPT Presentation

1 / 44
About This Presentation
Title:

Kees van Deemter

Description:

A pathological grammar of adjective order: NP the N(w) ... assume: one noun, many adjectives (not that it's easy to enforce this) Realization solution: ... – PowerPoint PPT presentation

Number of Views:47
Avg rating:3.0/5.0
Slides: 45
Provided by: matthe255
Category:
Tags: adjectives | deemter | kees | van

less

Transcript and Presenter's Notes

Title: Kees van Deemter


1
Formal IssuesinNatural Language Generation
  • Lecture 5
  • Stone, Doran,Webber, Bleam Palmer

2
GRE and surface realization
  • Arguably, GRE uses a grammar.
  • Parameters such as the preference order on
    properties reflect knowledge of how to
    communicate effectively.
  • Decisions about usefulness or completeness of a
    referring expression reflect beliefs about
    utterance interpretation.
  • Maybe this is a good idea for NLG generally.

3
GRE and surface realization
  • But weve thought GRE outputs semantics

referent furniture886 type desk status
definite color brown origin sweden
4
GRE and surface realization
  • We also need to link this up with surface form
  • the brown Swedish desk
  • Note not
  • ?the Swedish brown desk

5
Todays initial observations
  • Its hard to do realization on its own
  • mapping from semantics to surface structure.
  • Its easy to combine GRE and realization
  • because GRE is grammatical reasoning!
  • if you have a good representation for syntax.

6
Why its hard to do realization
  • A pathological grammar of adjective order
  • NP ? the N(w).
  • N(w) ? w N(w?) if w is an adjective and wRw?.
  • N(w) ? w if w is a noun.

7
Syntax with this grammar
  • Derivation of example

NP
N(brown)
N(Swedish)
N(desk)
the brown Swedish desk
Requires brown R Swedish, Swedish R desk
8
Realization, formally
  • You start with k properties.
  • Each property can be realized lexically.
  • assume one noun, many adjectives
  • (not that its easy to enforce this)
  • Realization solution
  • NP which realizes each property exactly once.

9
Quick formal analysis
  • View problem graph-theoretically
  • k words, corresponding to vertices in a graph
  • R is a graph on the k words
  • Surface structure is a Hamiltonian path
  • (which visits each vertex exactly once)
  • through R.
  • This is a famous NP complete problem
  • So surface realization itself is intractable!

10
Moral of the example
  • Semantics underdetermines syntactic relations.
  • Here, semantics underdetermines syntactic
    relations of adjectives to one another and to the
    head.
  • Searching for the correspondence is hard.
  • See also Brew 92, Koller and Striegnitz 02.

11
Todays initial observations
  • Its hard to do realization on its own
  • mapping from semantics to surface structure.
  • Its easy to combine GRE and realization
  • because GRE is grammatical reasoning!
  • if you have a good representation for syntax.

12
Syntactic processing for GRE
  • Lexicalization
  • Steps of grammatical derivation correspond to
    meaningful choices in NLG.
  • E.g., steps of grammar are synched with steps of
    adding a property to a description.

13
Syntactic processing for GRE
  • Key ideas lexicalization, plus
  • Flat dependency structure (adjs modify noun)
  • Hierarchical representation of word-order

NP
N(size)
N(color)
N(origin)
N(material)
the
desk
14
Syntactic processing for GRE
  • Other syntactic lexical entries

N(origin)
N(color)
Adj
Adj
Swedish
brown
15
Describing syntactic combination
  • Operation of combination 1 Substitution
  • NP

NP
NP
N(size)
N(size)
N(color)
N(color)
N(origin)
N(origin)
N(material)
N(material)
the
desk
the
desk
16
Describing syntactic combination
  • Operation of combination 2 Sister adjunction

NP
NP
N(color)
N(size)
N(size)
Adj
N(color)
N(color)
brown
N(origin)
N(origin)
Adj
N(material)
N(material)
the
desk
the
desk
brown
17
Abstracting syntax
  • Tree rewriting
  • Each lexical item is associated with a
    structure.
  • You have a starting structure.
  • You have ways of combining two structures
    together.

18
Abstracting syntax
  • Derivation tree
  • records elements and how they are combined

the desk
brown (s.a. _at_ color)
Swedish (s.a. _at_ origin)
19
An extended incremental algorithm
  • r individual to be described
  • P lexicon of entries, in preference order
  • P is an individual entry
  • sem(P) is a property or set of entries from the
    context
  • syn(P) is a syntactic element
  • L surface syntax of description

20
Extended incremental algorithm
  • L NP?
  • C Domain
  • For each P ?P do
  • If r ? sem(P) C ? sem(P)
  • Then do
  • L add(syn(P), L)
  • C C ? sem(P)
  • If C r then return L
  • Return failure

21
Observations
  • Why use tree-rewriting - not,e.g. CFG derivation?
  • NP ? the N(w).
  • N(w) ? w N(w?) if w is an adjective and wRw?.
  • N(w) ? w if w is a noun.
  • CFG derivation forces you to select properties in
    the surface word-order.

22
Observations
  • Tree-rewriting frees word-order from choice-order.

NP
NP
NP
N(size)
N(size)
N(size)
the
the
the
N(color)
N(color)
?
N(color)
?
N(origin)
N(origin)
Adj
N(origin)
Adj
N(material)
N(material)
Adj
N(material)
desk
brown
desk
brown
Swedish
desk
23
Observations
  • Tree-rewriting frees word-order from choice-order.

NP
NP
NP
N(size)
N(size)
N(size)
the
the
the
N(color)
N(color)
?
N(color)
?
N(origin)
N(origin)
N(origin)
Adj
N(material)
Adj
N(material)
Adj
N(material)
desk
Swedish
desk
brown
Swedish
desk
24
This is reflected in derivation tree
  • Derivation tree
  • records elements and how they are combined

the desk
brown (s.a. _at_ color)
Swedish (s.a. _at_ origin)
25
Formal results
  • Logical completeness.
  • If theres a flat derivation tree for an NP that
    identifies referent r,
  • Then the incremental algorithm finds it.
  • But
  • Sensible combinations of properties may not
    yield surface NPs.
  • Hierarchical derivation trees may require
    lookahead in usefulness check.

26
Formal results
  • Computational complexity
  • Nothing changes we just add properties, one
    after another

27
Now, though, were choosing specific lexical
entries
maybe these lexical items express the same
property
NP
NP
vs
N(departure)
N(departure)
the
the
N(destination)
Adj
N(destination)
Adj
N
N
N(stops)
N(stops)
335
Trenton
1535
Trenton
express
express
28
What motivates these choices?
  • Use
  • in 12-hour time context
  • Use
  • in 24-hour time context

N(departure)
Adj
335
N(departure)
Adj
1535
29
Need to extend grammar again
  • P lexicon of entries, in preference order
  • P is an individual entry
  • sem(P) is a property or set of entries from the
    context
  • syn(P) is a syntactic element
  • prags(P) is a test which the context must satisfy
    for the entry to be appropriate

30
Need to extend grammar again
  • For example
  • syn
  • sem departure(x, 1535)
  • prags twentyfourhourtime

N(departure)
Adj
1535
31
Extended incremental algorithm
  • L NP?
  • C Domain
  • For each P ?P do
  • If r ? sem(P) C ? sem(P) prags(P) is true
  • Then do
  • L add(syn(P), L)
  • C C ? sem(P)
  • If C r then return L
  • Return failure

32
DiscussionWhat does this entry do?
syn sem thing(x) prags in-focus(x)
NP
it
33
Suggestion find best value
  • Given
  • A set of entries that combine syntactically with
    L in the same way
  • Related by semantic generality and pragmatic
    specificity.
  • Current distractors
  • Take entries that remove the most distractors
  • Of those, take the most semantically general
  • Of those, take the most pragmatically specific

34
Extended incremental algorithm
  • L NP? C Domain
  • Repeat
  • Choices P add(syn(P), L) at next node
    r ? sem(P) prags(P) is true
  • P find best value(Choices)
  • L add(syn(P), L)
  • C C ? sem(P)
  • If C r then return L
  • Return failure

35
What is generation anyway?
  • Generation is intentional (or rational) action
  • thats why Grices maxims apply, for example.
  • You have a goal
  • You build a plan to achieve it
  • ( achieve it economically in a recognizable way)
  • You carry out the plan

36
In GRE
  • The goal is for hearer to know the identity of r
  • (in general g)
  • The plan will be to utter some NP U
  • such that the interpretation of U identifies r
  • (in general c ?u ? c?g)
  • Carrying out the plan means realizing this
    utterance.

37
In other words
  • GRE amounts to a process of deliberation.
  • Adding a property to L incrementally is like
    committing to an action.
  • These commitments are called intentions.
  • Incrementality is characteristic of intentions
    though in general intentions are open to
    revision.
  • Note this connects with belief-desire-intention
    models of bounded rationality.

38
GRE as (BDI) rational agency
  • L NP ? // Initial plan
  • C Domain // Interpretation
  • while (P FindBest(P, C, L)) //
    Deliberation
  • L add(syn(P), L) // Adopt new intention
  • C C ? sem(P) // Update interpretation
  • if C r return L // Goal satisfied
  • fail

39
NLG as (BDI) rational agency
  • L X ?
  • C Initial Interpretation
  • while (P FindBest(P, C, L))
  • L AddSyntax(syn(P), L)
  • C AddInterpretation(sem(P), C)
  • if GoalSatisfied(C) return L
  • fail

40
Conclusionsfor NLG researchers
  • Its worth asking (and answering) formal
    questions about NLG.
  • Questions of logical completeness can a
    generator express everything it ought to?
  • Questions of computational complexity is the
    cost of a generation algorithm worth the results?

41
Conclusionsfor linguists
  • NLG offers a precise perspective on questions of
    language use.
  • For example, whats the best way of
    communicating some message?
  • NLG as opposed to other perspectives gives
    more complete, smaller-scale models.

42
Conclusionsfor AI in general
  • NLG does force us to characterize and implement
    representations inference for practical
    interactive systems
  • Good motivation for computational semantics.
  • Meaty problems like logical form equivalence.
  • Many connections and possibilities for
    implementation (graphs, CSPs, circuit
    optimization, data mining,)

43
Open Problems
  • Sets and salience in REs.
  • Generating parallel REs.
  • Theoretical and empirical measures of
    quality/utility for REs.
  • Avoiding ambiguity in REs.
  • Any problem in RE generalizes to one in NLG.

44
Followup information
  • Course web page
  • http//www.itri.brighton.ac.uk/home/
  • Kees.van.Deemter/esslli-notes.html
  • downloadable papers
  • final lecture notes
  • papers weve talked about
  • links (recent/upcoming events, siggen, sigsem)
  • by Monday August 26.
Write a Comment
User Comments (0)
About PowerShow.com