Title: Learning%20the%20Structure%20of%20Task-Oriented%20Conversations%20from%20the%20Corpus%20of%20In-Domain%20Dialogs
1Learning the Structure of Task-Oriented
Conversations from the Corpus of In-Domain
Dialogs
- Ph.D. Thesis Defense
- Ananlada Chotimongkol
- Carnegie Mellon University, 18th December 2007
- Thesis Committee
- Alexander Rudnicky (Chair)
- William Cohen
- Carolyn Penstein Rosé
- Gokhan Tur (SRI International)
2Outline
- Introduction
- Structure of task-oriented conversations
- Machine learning approaches
- Conclusion
3A spoken dialog system
problem dialog structure learning approaches
conclusion
When would you like to leave?
I would like to fly to Seattle tomorrow.
Domain Knowledge
tasks, steps, domain keywords
Speech Synthesizer
Speech Recognizer
Natural Language Generator
Natural Language Understanding
Dialog Manager
4Problems in acquiring domain knowledge
problem dialog structure learning approaches
conclusion
Domain Knowledge (tasks, steps, domain keywords)
- Problems
- Require domain expertise
- Subjective
- May miss some cases (Yankelovich, 1997)
- Problems
- Require domain expertise
- Subjective
- May miss some cases
- Time consuming (Bangalore et al., 2006)
example dialogs
5Task-oriented dialog
problem dialog structure learning approaches
conclusion
- Observable structure
- Reflect domain information
- Observable -gt learnable?
Client I'D LIKE TO FLY TO HOUSTON TEXAS Agent
AND DEPARTING PITTSBURGH ON WHAT DATE
? Client DEPARTING ON FEBRUARY
TWENTIETH ... Agent DO YOU NEED A CAR
? Client YEAH Agent THE LEAST EXPENSIVE
RATE I HAVE WOULD BE WITH THRIFTY RENTAL CAR
FOR TWENTY THREE NINETY A DAY Client
OKAY Agent WOULD YOU LIKE ME TO BOOK THAT CAR
FOR YOU ? Client YES ... Agent OKAY AND
WOULD YOU NEED A HOTEL WHILE YOU'RE IN HOUSTON
? Client YES Agent AND WHERE AT IN HOUSTON
? Client /UM/ DOWNTOWN Agent OKAY Agent
DID YOU HAVE A HOTEL PREFERENCE ? ...
Client I'D LIKE TO FLY TO HOUSTON TEXAS Agent
AND DEPARTING PITTSBURGH ON WHAT DATE
? Client DEPARTING ON FEBRUARY
TWENTIETH ... Agent DO YOU NEED A CAR
? Client YEAH Agent THE LEAST EXPENSIVE
RATE I HAVE WOULD BE WITH THRIFTY RENTAL CAR
FOR TWENTY THREE NINETY A DAY Client
OKAY Agent WOULD YOU LIKE ME TO BOOK THAT CAR
FOR YOU ? Client YES ... Agent OKAY AND
WOULD YOU NEED A HOTEL WHILE YOU'RE IN HOUSTON
? Client YES Agent AND WHERE AT IN HOUSTON
? Client /UM/ DOWNTOWN Agent OKAY Agent
DID YOU HAVE A HOTEL PREFERENCE ? ...
6Proposed solution
problem dialog structure learning approaches
conclusion
Domain Knowledge (tasks, steps, domain keywords)
example dialogs
My Thesis
7Learning system output
problem dialog structure learning approaches
conclusion
8Thesis statement
problem dialog structure learning approaches
conclusion
Investigate how to infer domain-specific
information required to build a task-oriented
dialog system from a corpus of in-domain
conversations through an unsupervised
learning approach
9Thesis scope (1)
problem dialog structure learning approaches
conclusion
Investigate how to infer domain-specific
information required to build a task-oriented
dialog system from a corpus of in-domain
conversations through an unsupervised learning
approach
- What to learn domain-specific information in a
task-oriented dialog - A list of tasks and their decompositions
- (travel reservation flight, car, hotel)
- Domain keywords
- (airline, city name, date)
10Thesis scope (2)
problem dialog structure learning approaches
conclusion
Investigate how to infer domain-specific
information required to build a task-oriented
dialog system from a corpus of in-domain
conversations through an unsupervised
learning approach
- Resources a corpus of in-domain conversations
- Recorded human-human conversations are already
available
11Thesis scope (3)
problem dialog structure learning approaches
conclusion
Investigate how to infer domain-specific
information required to build a task-oriented
dialog system from a corpus of in-domain
conversations through an unsupervised
learning approach
- Learning approach unsupervised learning
- No training data available for a new domain
- Annotating data is time consuming
12Proposed approach
problem dialog structure learning approaches
conclusion
Investigate how to infer domain-specific
information required to build a task-oriented
dialog system from a corpus of in-domain
conversations through an unsupervised
learning approach
- 2 research problems
- Specify a suitable domain-specific information
representation - Develop a learning approach that infers domain
information captured by this representation from
human-human dialogs
13Outline
- Introduction
- Structure of task-oriented conversations
- Properties of a suitable dialog structure
- Form-based dialog structure representation
- Evaluation
- Machine learning approaches
- Conclusion
14Properties of a desired dialog structure
problem dialog structure properties learning
approaches conclusion
- Sufficiency
- Capture all domain-specific information required
to build a task-oriented dialog system - Generality (domain-independent)
- Able to describe task-oriented dialogs in
dissimilar domains and types - Learnability
- Can be identified by an unsupervised machine
learning algorithm
15Domain-specific informationin task-oriented
dialogs
problem dialog structure properties
learning approaches conclusion
- A list of tasks and their decompositions
- Ex travel reservation flight car hotel
- A compositional structure of a dialog based on
the characteristics of a task - Domain keywords
- Ex airline, city name, date
- The actual content of a dialog
16Existing discourse structures
problem dialog structure properties
learning approaches conclusion
Discourse structure Sufficiency Generality Learnability
Segmented Discourse Representation Theory (Asher, 1993) Focus on meaning not actual entities ? ?
Grosz and Sidners Theory (Grosz and Sidner, 1986) Doesnt model domain keywords ? unsupervised?
DAMSL extension (Hardy et al., 2003) Doesnt model a compositional structure ? unsupervised?
A plan-based model (Cohen and Perrault, 1979) ? ? unsupervised?
17Form-based dialog structure representation
problem dialog structure form-based
learning approaches conclusion
- Based on a notion of form (Ferrieux and Sadek,
1994) - A data representation used in the form-based
dialog system architecture - Focus only on concrete information
- Can be observed directly from in-domain
conversations
18Form-based representation components
problem dialog structure form-based
learning approaches conclusion
- Consists of 3 components
- Task
- Sub-task
- Concept
19Form-based representation components
- Task
- A subset of a dialog that has a specific goal
Client I'D LIKE TO FLY TO HOUSTON TEXAS Agent
AND DEPARTING PITTSBURGH ON WHAT DATE
? Client DEPARTING ON FEBRUARY
TWENTIETH ... Agent DO YOU NEED A CAR
? Client YEAH Agent THE LEAST EXPENSIVE
RATE I HAVE WOULD BE WITH THRIFTY RENTAL CAR
FOR TWENTY THREE NINETY A DAY Client
OKAY Agent WOULD YOU LIKE ME TO BOOK THAT CAR
FOR YOU ? Client YES ... Agent OKAY AND
WOULD YOU NEED A HOTEL WHILE YOU'RE IN HOUSTON
? Client YES Agent AND WHERE AT IN HOUSTON
? Client /UM/ DOWNTOWN Agent OKAY Agent
DID YOU HAVE A HOTEL PREFERENCE ? ...
20Form-based representation components
- Sub-task
- A step in a task that contributes toward the goal
- Contains sufficient information to execute a
domain action
Client I'D LIKE TO FLY TO HOUSTON TEXAS Agent
AND DEPARTING PITTSBURGH ON WHAT DATE
? Client DEPARTING ON FEBRUARY
TWENTIETH ... Agent DO YOU NEED A CAR
? Client YEAH Agent THE LEAST EXPENSIVE
RATE I HAVE WOULD BE WITH THRIFTY RENTAL CAR
FOR TWENTY THREE NINETY A DAY Client
OKAY Agent WOULD YOU LIKE ME TO BOOK THAT CAR
FOR YOU ? Client YES ... Agent OKAY AND
WOULD YOU NEED A HOTEL WHILE YOU'RE IN HOUSTON
? Client YES Agent AND WHERE AT IN HOUSTON
? Client /UM/ DOWNTOWN Agent OKAY Agent
DID YOU HAVE A HOTEL PREFERENCE ? ...
21Form-based representation components
- Concept (domain keywords)
- A piece of information required to perform an
action
Client I'D LIKE TO FLY TO HOUSTON TEXAS Agent
AND DEPARTING PITTSBURGH ON WHAT DATE
? Client DEPARTING ON FEBRUARY
TWENTIETH ... Agent DO YOU NEED A CAR
? Client YEAH Agent THE LEAST EXPENSIVE
RATE I HAVE WOULD BE WITH THRIFTY RENTAL CAR
FOR TWENTY THREE NINETY A DAY Client
OKAY Agent WOULD YOU LIKE ME TO BOOK THAT CAR
FOR YOU ? Client YES ... Agent OKAY AND
WOULD YOU NEED A HOTEL WHILE YOU'RE IN HOUSTON
? Client YES Agent AND WHERE AT IN HOUSTON
? Client /UM/ DOWNTOWN Agent OKAY Agent
DID YOU HAVE A HOTEL PREFERENCE ? ...
22Data representation
problem dialog structure form-based
learning approaches conclusion
- Represented by a form
- A repository of related pieces of information
necessary for performing an action
23Data representation
- Form a repository of related pieces of
information - Sub-task contains sufficient information to
execute a domain action ? a form
Client I'D LIKE TO FLY TO HOUSTON TEXAS Agent
AND DEPARTING PITTSBURGH ON WHAT DATE
? Client DEPARTING ON FEBRUARY
TWENTIETH ... Agent DO YOU NEED A CAR
? Client YEAH Agent THE LEAST EXPENSIVE
RATE I HAVE WOULD BE WITH THRIFTY RENTAL CAR
FOR TWENTY THREE NINETY A DAY Client
OKAY Agent WOULD YOU LIKE ME TO BOOK THAT CAR
FOR YOU ? Client YES ... Agent OKAY AND
WOULD YOU NEED A HOTEL WHILE YOU'RE IN HOUSTON
? Client YES Agent AND WHERE AT IN HOUSTON
? Client /UM/ DOWNTOWN Agent OKAY Agent
DID YOU HAVE A HOTEL PREFERENCE ? ...
24Data representation
- Form a repository of related pieces of
information - Task a subset of a dialog that has a specific
goal - ? a set of forms
Client I'D LIKE TO FLY TO HOUSTON TEXAS Agent
AND DEPARTING PITTSBURGH ON WHAT DATE
? Client DEPARTING ON FEBRUARY
TWENTIETH ... Agent DO YOU NEED A CAR
? Client YEAH Agent THE LEAST EXPENSIVE
RATE I HAVE WOULD BE WITH THRIFTY RENTAL CAR
FOR TWENTY THREE NINETY A DAY Client
OKAY Agent WOULD YOU LIKE ME TO BOOK THAT CAR
FOR YOU ? Client YES ... Agent OKAY AND
WOULD YOU NEED A HOTEL WHILE YOU'RE IN HOUSTON
? Client YES Agent AND WHERE AT IN HOUSTON
? Client /UM/ DOWNTOWN Agent OKAY Agent
DID YOU HAVE A HOTEL PREFERENCE ? ...
25Data representation
- Form a repository of related pieces of
information - Concept a piece of information required to
perform an action ? a slot
Client I'D LIKE TO FLY TO HOUSTON TEXAS Agent
AND DEPARTING PITTSBURGH ON WHAT DATE
? Client DEPARTING ON FEBRUARY
TWENTIETH ... Agent DO YOU NEED A CAR
? Client YEAH Agent THE LEAST EXPENSIVE
RATE I HAVE WOULD BE WITH THRIFTY RENTAL CAR
FOR TWENTY THREE NINETY A DAY Client
OKAY Agent WOULD YOU LIKE ME TO BOOK THAT CAR
FOR YOU ? Client YES ... Agent OKAY AND
WOULD YOU NEED A HOTEL WHILE YOU'RE IN HOUSTON
? Client YES Agent AND WHERE AT IN HOUSTON
? Client /UM/ DOWNTOWN Agent OKAY Agent
DID YOU HAVE A HOTEL PREFERENCE ? ...
26Form-based representation properties
problem dialog structure form-based
learning approaches conclusion
- Sufficiency
- The form is already used in a form-based dialog
system - Philips train timetable system (Aust et al.,
1995) - CMU Communicator system (Rudnicky et al., 1999)
- Generality (domain-independent)
- A broader interpretation of the form is provided
- The analysis of six dissimilar domains
- Learnability
- Components are observable directly from a dialog
- (by human) annotation scheme reliability
- (by machine) the accuracy of the domain
information learned by the proposed approaches
27Outline
- Introduction
- Structure of task-oriented conversations
- Properties of a suitable dialog structure
- Form-based dialog structure representation
- Evaluation
- Dialog structure analysis (generality)
- Annotation experiment (human learnability)
- Machine learning approaches
- Conclusion
28Dialog structure analysis
problem dialog structure analysis learning
approaches conclusion
- Goal
- To verify that form-based representation can be
applied to dissimilar domains - Approach
- Analyze 6 task-oriented domains
- Air travel planning (information-accessing task)
- Bus schedule inquiry (information-accessing task)
- Map reading (problem-solving task)
- UAV flight simulation (command-and-control task)
- Meeting (personnel resource management)
- Tutoring (physics essay revising)
29Map reading domain
route giver
route follower
30Map reading domain(problem-solving task)
problem dialog structure analysis learning
approaches conclusion
- Task draw a route on a map
- Sub-task draw a segment of a route
- Concepts
- StartLocation White_Mountain, Machete,
- Direction down, left,
- Distance a couple of centimeters, an inch,
- Sub-task ground a landmark
- Concepts
- LandmarkName White_Mountain, Machete,
- Location below the start,
31Dialog structure analysis (map reading domain)
GIVER 1 okay ... ehm ... right, you have the
start? FOLLOWER 2 yeah. GIVER 3 right,
below the start do you have ... er like a
missionary camp? FOLLOWER 4 yeah. GIVER 5
okay, well ... if you take it from the start
just run ... horizontally. FOLLOWER 6
uh-huh. GIVER 7 eh to the left for about an
inch. FOLLOWER 8 right. GIVER 9 and then
go down along the side of the missionary
camp. FOLLOWER 10 uh-huh. GIVER 11
'til you're about an inch ... above the bottom of
the map. FOLLOWER 12 right. GIVER 13
then you need to go straight along for about
'til about ...
GIVER 1 okay ... ehm ... right, you have the
start? FOLLOWER 2 yeah. (action (implicit)
define_a_landmark) GIVER 3 right, below the
start do you have ... er like a missionary
camp? FOLLOWER 4 yeah. (action
define_a_landmark) GIVER 5 okay, well ... if
you take it from the start just run ...
horizontally. FOLLOWER 6 uh-huh. GIVER 7
eh to the left for about an inch. FOLLOWER 8
right. (action draw_a_segment) GIVER 9 and
then go down along the side of the missionary
camp. FOLLOWER 10 uh-huh. GIVER 11
'til you're about an inch ... above the bottom of
the map. FOLLOWER 12 right. GIVER
13 then you need to go straight along for
about 'til about ...
32UAV flight simulation domain(command-and-control
task)
problem dialog structure analysis learning
approaches conclusion
- Task take photos of the targets
- Sub-task take a photo of each target
- Sub-subtask control a plane
- Concepts
- Altitude 2700, 3300,
- Speed 50 knots, 200 knots,
- Destination H-area, SSTE,
- Sub-subtask ground a landmark
- Concepts
- LandmarkName H-area, SSTE,
- LandmarkType target, waypoint
33Meeting domain
problem dialog structure analysis learning
approaches conclusion
- Task manage resources for a new employee
- Sub-task get a computer
- Concepts
- Type desktop, laptop,
- Brand IBM, Dell,
- Sub-task get office space
- Sub-task create an action item
- Concepts
- Description have a space,
- Person Hardware Expert, Building Expert,
- StartDate today,
- EndDate the fourteenth of december,
34Characteristics of form-based representation
problem dialog structure analysis learning
approaches conclusion
- Focus only on concrete information
- That is observable directly from in-domain
conversations - Describe a dialog with a simple model
- Pros
- Possible to be learned by an unsupervised
learning approach - Cons
- Cant capture information that is not clearly
expressed in a dialog - Omitted concept values
- Nevertheless, 93 of dialog content can be
accounted for - Cant model a complex dialog that has a dynamic
structure - A tutoring domain
- But it is good enough for many real world
applications
35Form-based representation properties(revisit)
problem dialog structure analysis learning
approaches conclusion
- Sufficiency
- The form is already used in a form-based dialog
system - Can account for 93 of dialog content
- Generality (domain-independent)
- A broader interpretation of the form
representation is provided - Can represent 5 out of 6 disparate domains
- Learnability
- Components are observable directly from a dialog
- (by human) annotation scheme reliability
- (by machine) the accuracy of the domain
information learned by the proposed approaches
36Annotation experiment
problem dialog structure annotation
experiment learning approaches conclusion
- Goal
- To verify that the form-based representation can
be understood and applied by other annotators - Approach
- Conduct an annotation experiment with non-expert
annotators - Evaluation
- Similarity between annotations
- Accuracy of annotations
37Challenges in annotation comparison
problem dialog structure annotation
experiment learning approaches conclusion
- Different tagsets may be used since annotators
have to design theirs own tagsets - Some differences are acceptable if they conform
to the guideline - Different dialog structure designs can generate
dialog systems with the same functionalities
Annotator 1 Annotator 2
ltNoOfStopgt -
ltDestinationCitygt ltDestinationLocationgtltCitygt
ltDategt ltDepartureDategt and ltArrivalDategt
38Cross-annotator correction
- Each annotator creates his/her own tagset and
then annotate dialogs - Each annotator critiques and corrects another
annotators work - Compare the original annotation with the
corrected one
39Annotation experiment
problem dialog structure annotation
experiment learning approaches conclusion
- 2 domains
- Air travel planning domain (information-accessing
task) - Map reading domain (problem-solving task)
- 4 subjects in each domain
- People who are likely to use the form-based
representation in the future - Each subject has to
- Design a tagset and annotate the structure of
dialogs - Critique other subjects annotation on the same
set of dialogs
40Evaluation metrics
- Annotation similarity
- Acceptability is the degree to which an original
annotation is acceptable to a corrector - Annotation accuracy
- Accuracy is the degree to which a subjects
annotation is acceptable to an expert
41Annotation results
problem dialog structure annotation
experiment learning approaches conclusion
Concept Annotation Air Travel Map Reading
acceptability 0.96 0.95
accuracy 0.97 0.89
Task/subtask Annotation Air Travel Map Reading
acceptability 0.81 0.84
accuracy 0.90 0.65
- High acceptability and accuracy
- Except task/sub-task accuracy in map reading
domain - Concepts can be annotated more reliably than
tasks and sub-tasks - Smaller units
- Have to be communicated clearly
42Form-based representation properties(revisit)
problem dialog structure annotation
experiment learning approaches conclusion
- Sufficiency
- The form is already used in a form-based dialog
system - Can account for 93 of dialog content
- Generality (domain-independent)
- A broader interpretation of the form
representation is provided - Can represent 5 out of 6 disparate domains
- Learnability
- Components are observable directly from a dialog
- Can be applied reliably by other annotators in
most of the cases - (by machine) the accuracy of the domain
information learned by the proposed approaches
43Outline
- Introduction
- Structure of task-oriented conversations
- Machine learning approaches
- Conclusion
44Overview of learning approaches
problem dialog structure learning approaches
conclusion
- Divide into 2 sub-problems
- Concept identification
- What are the concepts?
- What are their members?
- Form identification
- What are the forms?
- What are the slots (concepts) in each form?
- Use unsupervised learning approaches
- Acquisition (not recognition) problem
45Learning example
problem dialog structure learning approaches
conclusion
Client I'D LIKE TO FLY TO HOUSTON TEXAS Agent
AND DEPARTING PITTSBURGH ON WHAT DATE
? Client DEPARTING ON FEBRUARY
TWENTIETH ... Agent DO YOU NEED A CAR
? Client YEAH Agent THE LEAST EXPENSIVE
RATE I HAVE WOULD BE WITH THRIFTY RENTAL CAR
FOR TWENTY THREE NINETY A DAY Client
OKAY Agent WOULD YOU LIKE ME TO BOOK THAT CAR
FOR YOU ? Client YES ... Agent OKAY AND
WOULD YOU NEED A HOTEL WHILE YOU'RE IN HOUSTON
? Client YES Agent AND WHERE AT IN HOUSTON
? Client /UM/ DOWNTOWN Agent OKAY Agent
DID YOU HAVE A HOTEL PREFERENCE ? ...
Client I'D LIKE TO FLY TO HOUSTON TEXAS Agent
AND DEPARTING PITTSBURGH ON WHAT DATE
? Client DEPARTING ON FEBRUARY
TWENTIETH ... Agent DO YOU NEED A CAR
? Client YEAH Agent THE LEAST EXPENSIVE
RATE I HAVE WOULD BE WITH THRIFTY RENTAL CAR
FOR TWENTY THREE NINETY A DAY Client
OKAY Agent WOULD YOU LIKE ME TO BOOK THAT CAR
FOR YOU ? Client YES ... Agent OKAY AND
WOULD YOU NEED A HOTEL WHILE YOU'RE IN HOUSTON
? Client YES Agent AND WHERE AT IN HOUSTON
? Client /UM/ DOWNTOWN Agent OKAY Agent
DID YOU HAVE A HOTEL PREFERENCE ? ...
46Outline
- Introduction
- Structure of task-oriented conversations
- Machine learning approaches
- Concept identification
- Form identification
- Conclusion
47Concept identification
problem dialog structure learning approaches
concept identification conclusion
- Goal Identify domain concepts and their members
- CityPittsburgh, Boston, Austin,
- MonthJanuary, February, March,
- Approach word clustering algorithm
- Identify concept words and group the similar ones
into the same cluster
48Word clustering algorithms
problem dialog structure learning approaches
concept identification conclusion
- Use word co-occurrences statistics
- Mutual information (MI-based)
- Kullback-Liebler distance (KL-based)
- Iterative algorithms need a stopping criteria
- Use information that is available during the
clustering process - Mutual information (MI-based)
- Distance between clusters (KL-based)
- Number of clusters
49Clustering evaluation
problem dialog structure learning approaches
concept identification conclusion
- Allow more than one cluster to represent a
concept - To discover as many concept words as possible
- However, the clustering result that doesnt
contain splited concepts is preferred - Quality score (QS) harmonic mean of
- Precision (purity)
- Recall (completeness)
- Singularity Score (SS)
- SS of conceptj
50Concept clustering results
problem dialog structure learning approaches
concept identification conclusion
Algorithm Precision Recall SS QS MaxQS
MI-based 0.78 0.43 0.77 0.61 0.68
KL-based 0.86 0.60 0.70 0.70 0.71
- Domain concepts can be identified with acceptable
accuracy - Example clusters
- GATWICK, CINCINNATI, PHILADELPHIA, L.A.,
ATLANTA - HERTZ, BUDGET, THRIFTY
- Low recall for infrequent concepts
- An automatic stopping criterion yields close to
optimal results
51Outline
- Introduction
- Structure of task-oriented conversations
- Machine learning approaches
- Concept identification
- Form identification
- Conclusion
52Form Identification
problem dialog structure learning approaches
form identification conclusion
- Goal determine different types of forms and
their associated slots - Approach
- Segment a dialog into a sequence of sub-tasks
- Dialog segmentation
- Group the sub-tasks that associate with the same
form type into a cluster - Sub-task clustering
- Identify a set of slots associated with each form
type - Slot extraction
53Step1 dialog segmentation
problem dialog structure learning approaches
form identification conclusion
- Goal segment a dialog into a sequence of
sub-tasks - Equivalent to identify sub-task boundaries
- Approach
- TextTiling algorithm (Hearst, 1997)
- Based on lexical cohesion assumption (local
context) - HMM-based segmentation algorithm
- Based on recurring patterns (global context)
- HMM states topics (sub-tasks)
- Transition probability probabilities of topic
shifts - Emission probability a state-specific language
model
54Modeling HMM states
- HMM states topics (sub-tasks)
- Induced by clustering reference topics (Tür et
al., 2001) - Need annotated data
- Utterance-based HMM (Barzilay and Lee, 2004)
- Some utterances are very short
- Induced by clustering predicted segments from
TextTiling
55Modifications for fine-grained segments in spoken
dialogs
problem dialog structure learning approaches
form identification conclusion
- Average segment length
- Air travel domain 84 words
- Map reading domain 55 words
- (WSJ 428, Broadcast News 996)
- Modifications include
- A data-driven stop word list
- Reflect the characteristics of spoken dialogs
- A distance weight
- Higher weight for the context closer to candidate
boundary
56Dialog segmentation experiment
problem dialog structure learning approaches
form identification conclusion
- Evaluation metrics
- Pk (Beeferman et al., 1999)
- Probabilistic error metric
- Sensitive to the value of k
- Concept-based F-measure (C. F-1)
- F-measure (or F-1) is a harmonic mean of
precision and recall - Count a near miss as a match if there is no
concept in between - Incorporate concept information in word token
representation - A concept label its value -gt
Airlinenorthwest - A concept label -gt Airline
57TextTiling results
problem dialog structure learning approaches
form identification conclusion
Algorithm Air Travel Air Travel Map Reading Map Reading
Algorithm Pk C. F-1 Pk C. F-1
TextTiling (baseline) 0.387 0.621 0.412 0.396
TextTiling (augmented) 0.371 0.712 0.384 0.464
- Augmented TextTiling is significantly better than
the baseline
58HMM-based segmentation results
problem dialog structure learning approaches
form identification conclusion
Algorithm Air Travel Air Travel Map Reading Map Reading
Algorithm Pk C. F-1 Pk C. F-1
HMM-based (utterance) 0.398 0.624 0.392 0.436
HMM-based (segment) 0.385 0.698 0.355 0.507
HMM-based (segment label) 0.386 0.706 0.250 0.686
TextTiling (augmented) 0.371 0.712 0.384 0.464
- Inducing HMM states from predicted segments is
better than inducing from utterances - Abstract concept representation yields better
results - Especially on map reading domain
- HMM-based is significantly better than TextTiling
on map reading domain
59Segmentation error analysis
problem dialog structure learning approaches
form identification conclusion
- TextTiling algorithm performs better on
consecutive sub-tasks of the same type - HMM-based algorithm performs better on very
fine-grained segments (only 2-3 utterances long) - Map reading domain
60Step2 sub-task clustering
problem dialog structure learning approaches
form identification conclusion
- Approach
- Bisecting K-mean clustering algorithm
- Incorporate concept information in word token
representation - Evaluation metrics
- Similar to concept clustering
61Sub-task clustering results
problem dialog structure learning approaches
form identification conclusion
Concept Word Representation Air Travel Map Reading
concept label value (oracle segment) 0.738 0.791
concept label value 0.577 0.675
concept label 0.601 0.823
- Inaccurate segment boundaries affect clustering
performance - But dont affect frequent sub-tasks much
- Missing boundaries are more problematic than
false alarms - Abstract concept representation yields better
results - More improvement in the map reading domain
- Even better than using reference segments
- Appropriate feature representation is better than
accurate segment boundaries
62Step3 Slot extraction
problem dialog structure learning approaches
form identification conclusion
- Goal
- Identify a set of slots associated with each form
type - Approach
- Analyze concepts contained in each cluster
63Slot extraction results
problem dialog structure learning approaches
form identification conclusion
- Concepts are sorted by frequency
64Outline
- Introduction
- Structure of task-oriented conversations
- Machine learning approaches
- Concept identification and clustering
- Form identification
- Conclusion
65Form-based dialog structure representation
problem dialog structure learning approaches
conclusion
- Forms are a suitable domain-specific information
representation according to these criteria - Sufficiency
- Can account for 93 of dialog content
- Generality (domain-independent)
- A broader interpretation of the form
representation is provided - Can represent 5 out of 6 disparate domains
- Learnability
- (human) can be applied reliably by other
annotators in most of the cases - (machine) can be identified with acceptable
accuracy using unsupervised machine learning
approaches
66Unsupervised learning approaches for inferring
domain information
- Require some modifications in order to learn the
structure of a spoken dialog - Can identify components in form-based
representation with acceptable accuracy - Concept accuracy, QS 0.70
- Sub-task boundary accuracy, F-1 0.71 (air
travel), 0.69 (map
reading) - Form type accuracy, QS 0.60 (air travel),
0.82 (map
reading) - Can learn with inaccurate information
- If the number of errors is moderate
- Propagated errors dont affect frequent
components much - Dialog structure acquisition doesnt require high
learning accuracy
67Conclusion
- To represent a dialog for a learning purpose we
based our representation on an observable
structure - This observable representation
- Can be generalize for various types of
task-oriented dialog - Can be understood and applied by different
annotators - Can be learned by unsupervised learning approach
- The result from this investigation can be apply
for - Acquiring domain knowledge in a new task
- Exploring the structure of a dialog
- Could potentially reduce human effort when
developing a new dialog system
68Thank you
69References (1)
- N. Asher. 1993. Reference to Abstract Objects in
Discourse. Dordrecht, the Netherlands Kluwer
Academic Publishers. - H. Aust, M. Oerder, F. Seide, and V. Steinbiss.
1995. The Philips automatic train timetable
information system. Speech Communication,
17(3-4)249-262. - S. Bangalore, G. D. Fabbrizio, and A. Stent.
2006. Learning the Structure of Task-Driven
Human-Human Dialogs. In Proceedings of COLING/ACL
2006. Sydney, Australia. - R. Barzilay and L. Lee. 2004. Catching the Drift
Probabilistic Content Models, with Applications
to Generation and Summarization. In HLT-NAACL
2004 Proceedings of the Main Conference, pp.
113-120. Boston, MA. - D. Beeferman, A. Berger, and J. Lafferty. 1999.
Statistical Models for Text Segmentation. Machine
Learning, 34(1-3)177-210. - P. R. Cohen and C. R. Perrault. 1979. Elements of
a plan-based theory of speech acts. Cognitive
Science, 3177-212. - A. Ferrieux and M. D. Sadek. 1994. An Efficient
Data-Driven Model for Cooperative Spoken
Dialogue. In Proceedings of ICSLP 1994. Yokohama,
Japan. - B. J. Grosz and C. L. Sidner. 1986. Attention,
intentions, and the structure of discourse.
Computational Linguistics, 12(3)175-204.
70References (2)
- H. Hardy, K. Baker, H. Bonneau-Maynard, L.
Devillers, S. Rosset, and T. Strzalkowski. 2003.
Semantic and Dialogic Annotation for Automated
Multilingual Customer Service. In Proceedings of
Eurospeech 2003. Geneva, Switzerland. - M. A. Hearst. 1997. TextTiling segmenting text
into multi-paragraph subtopic passages.
Computational Linguistics, 23(1)33-64. - W. C. Mann and S. A. Thompson. 1988. Rhetorical
Structure Theory Toward a functional theory of
text organization. Text, 8(3)243-281. - L. Polanyi. 1996. The Linguistic Structure of
Discourse, Technical Report CSLI-96-200. Stanford
CA, Center for the Study of Language and
Information, Stanford University. - A. I. Rudnicky, E. Thayer, P. Constantinides, C.
Tchou, R. Shern, K. Lenzo, X. W., and A. Oh.
1999. Creating natural dialogs in the Carnegie
Mellon Communicator system. In Proceedings of
Eurospeech 1999. Budapest, Hungary. - J. M. Sinclair and M. Coulthard. 1975. Towards an
analysis of Discourse The English used by
teachers and pupils Oxford University Press. - G. Tür, A. Stolcke, D. Hakkani-Tür, and E.
Shriberg. 2001. Integrating prosodic and lexical
cues for automatic topic segmentation.
Computational Linguistics, 27(1)31-57. - N. Yankelovich. 1997. Using Natural Dialogs as
the Basis for Speech Interface Design. In Susann
Luperfoy (Ed.), Automated Spoken Dialog Systems.
Cambridge, MA MIT Press.