Towards Interactive Training with an Avatarbased HumanComputer Interface - PowerPoint PPT Presentation

1 / 20
About This Presentation
Title:

Towards Interactive Training with an Avatarbased HumanComputer Interface

Description:

Towards Interactive Training with an Avatar-based. Human-Computer Interface ... Make an avatar that is visually recognizable as Dr. Schwarzkopf, both in facial ... – PowerPoint PPT presentation

Number of Views:79
Avg rating:3.0/5.0
Slides: 21
Provided by: Stephen694
Learn more at: https://www.evl.uic.edu
Category:

less

Transcript and Presenter's Notes

Title: Towards Interactive Training with an Avatarbased HumanComputer Interface


1
Towards Interactive Training with an
Avatar-basedHuman-Computer Interface
UCF
University of Central Florida University of
Illinois at Chicago
2
Topics
  • What we are doing
  • Why are we doing it
  • How we are doing it
  • How this relates to simulation and training
  • Results
  • Summary

2
3
What are we doing? briefly
  • Building a guru system with an avatar interface
    capable of communicating with humans via spoken
    language
  • Our current emphasis is on avatar interface
  • Lifelike in its appearance and dialog
  • Project funded by NSF to a collaborative research
    team at the University of Central Florida (UCF)
    and the University of Illinois at Chicago (UIC)

4
The What (continued)
  • The initial objective replace a specific human
    being Dr. Alex Schwarzkopf at the NSF
  • Founding and long-time director of the
    Industry/University Collaborative Research
    Centers (IUCRC) program
  • Recently retired
  • Follows a separate three-year research project at
    UCF to capture Dr. Schwarzkopfs knowledge about
    the I/UCRC and code it in an easily retrievable
    system (called AskAlex)

5
Dr. Alex Schwarzkopf
6
The AlexAvatar
7
Challenges
  • Facing several challenges in this project
  • Make an avatar that is visually recognizable as
    Dr. Schwarzkopf, both in facial features and in
    his mannerisms and tendencies.
  • Provide a natural language interface that
    understands the questions from the user.
  • Manage the dialog in a way that is natural as
    well as effective.
  • Interface the system to AskAlex that can answer
    questions posed by users on or about the I/UCRC.
  • Simulate his voice, choice of words and
    inflections as closely as possible.

8
Why are we doing this?
  • We seek to show the feasibility of effective
    interfaces that employ natural language and a
    recognizable human-like avatar.
  • Is it practical?
  • Do humans respond better to lifelike avatars than
    text-based interfaces?
  • We are trying to preserve the legacy of Dr.
    Schwarzkopf by not only preserving his knowledge
    but in many ways himself, too.

9
How are we doing this?
  • Two parts to the work
  • UIC charged with the visual elements of avatar
  • UCF charged with the communication and general
    intelligence of avatar

10
Creating a Virtual Alex
  • Vicon motion capture hardware

11
Creating a Virtual Alex (cont)
  • FaceGen software for expressions

12
Alex at Work
  • Alex sits at his desk at NSF, a familiar
    environment (sitting at desk in office wearing a
    suit).
  • Alex provides
  • information
  • by speaking and
  • showing textual
  • or graphical
  • information on
  • the TV screen
  • in his office.

13
Interacting with the User
  • Alex displayed on a 50 display
  • Shows upper body with room for him to gesture
  • Multiple microphones used so Alex can turn
    towards speaker
  • Blinking and other involuntary motions based on
    Alexs mannerisms
  • Responsive Avatar Framework
  • Skeleton Animation Synthesizer
  • Facial Expression Synthesizer
  • Lip Synchronizer

14
Speech Recognizer Architecture
Operational View
LifeLike Recognizer
ChantSR
MS SAPI 5.1
Dictation Mode
LifeLike Dialog System
Grammar Recognition
Layered Recognition Design
Dictation Mode Text Repository
Grammars XML
15
Dialog System Architecture
LifeLike Dialog System
LifeLike Recognizer
Knowledge Manager
Speech Disambiguator
Spell Check
Dictation String
Context
Phrase String
Semantics Check
Dataset
Disambiguated String
NSF User Data
Context
Context-based Dialogue Manager
Context Specific Knowledge
AskAlex Ontology
Context
Response String
General Knowledge
Context
LifeLike Speech Output
Response String
Dataset
Updated Data
16
Dialog Management
  • Semantic Disambiguator
  • Spelling and semantic check of SR input
  • Uses contextual matching processes
  • Knowledge Manager
  • 3 sources of knowledge General, Domain-specific,
    User Profile
  • Contextualized Knowledge Base
  • CxBR-based Dialog Manager
  • Goal recognition
  • Inference Engine determines state of conversation
    for contextual relevance
  • Goal management
  • Goal Stack keeps track of conversational goals
  • Agent actions
  • Context Topology dictates conversational output
    using Inference Engine and Goal Stack

17
Relation to ST
  • Perform functions done by humans
  • Mixed Reality Training, AAR, Concierges,
    Interrogator Training, Intelligent Tutoring
    Systems
  • Multimedia Integration
  • hybrid speech and pointing interface with user
  • Text, graphics, links to clarifying information
    or documents
  • Customizable environments support changing
    mission requirements
  • Characteristics (age, gender, appearance) can be
    varied independent of domain knowledge to match
    each trainee
  • Avatar can represent specific individual CO, DI
  • Surroundings in which avatar occupies can be
    updated

18
Video
  • ltinsert new video heregt

19
Initial Evaluation
  • Controlled experiment
  • 30 diverse students from UIC
  • study displayed ten 30-second life-size videos of
    AlexAvatar
  • Videos paired to compare/contrast
  • Maintain eye contact, Head motion, Body motion,
    Pre-recorded vs computer generated Voice
  • Baseline comparison to our previous avatar
    implementation from a year ago.
  • Test subjects preferred by a significant margin
  • Body movement compared to a still body
  • An avatar with purposeful motions incorporating
    motion-capture data from Alex
  • More detailed and realistic textures
  • No clear preference in pre-recorded vs.
    computer-generated voice

20
Summary
  • Currently on year two of three-year project
  • Current results are encouraging, several
    challenges still ahead
  • Focusing on
  • avatar handling conversation with more than one
    human,
  • Increasing capacity for understanding spoken
    language
  • Expanding capabilities of dialog system,
  • Evaluating effect of whiteboard feature.
  • Expect to finish second edition of avatar soon
    for demo in early January at NSF conference
  • Looking for partnerships with military
    applications
Write a Comment
User Comments (0)
About PowerShow.com