Multimodal Interfaces Robust interaction where graphical user interfaces fear to tread

Download Report

Transcript Multimodal Interfaces Robust interaction where graphical user interfaces fear to tread

Multimodal Interfaces
Robust interaction where graphical
user interfaces fear to tread
Philip R. Cohen
Professor and Co-Director
Center for Human-Computer Communication
Oregon Health and Science Univ.
http://www.cse.ogi.edu/CHCC
and
Natural Interaction Systems, LLC
Team Effort
Co-PI: Sharon Oviatt
Rajah Annamalai
Alex Arthur
Paulo Barthelmess
Rachel Coulston
Marisa Flecha-Garcia
Xiao Huang
Ed Kaiser
Sanjeev Kumar
Rebecca Lunsford
Richard Wesson
Multidisciplinary research
Multimodal Interaction
• Use of one or more natural communication
modalities—e.g. , Speech, gesture, sketch …
• Advantages over GUI and unimodal systems
– Easier to use; Less training
– Robust, flexible
– Preferred by users
– Faster, more efficient
– Supports new functionality
• Applies to many different environments and
form factors that challenge GUI, especially
mobile ones
Potential Application Areas
•
•
•
•
•
•
•
•
•
•
Architecture and Design
Geographical Information Systems
Emergency Operations
Field-based Operations
Mobile Computing and Telecommunications
Virtual/Augmented Reality
Pervasive/Ubiquitous Computing
Computer-Supported Collaborative Work
Education
Entertainment
Challenges for multimodal interface design
• More than 2 modes –e.g. spoken, gestural,
facial expression, gaze; various sensors
• Inputs are uncertain –vs. Keyboard/mouse
– Corrupted by noise
– Multiple people
• Recognition is probabilistic
• Meaning is ambiguous
Design for uncertainty
Approach
Gain robustness via
– Fusion of inputs from multiple modalities
– Using strengths of one mode to compensate for
weaknesses of others—design time and run time
– Avoiding/correcting errors
– Statistical architecture
– Confirmation
– Dialogue context
– Simplification of language in a multimodal context
– Output affecting/channeling input
Demo
Started with 50 & 100Mhz 486
Multimodal Architecture
Late MM Integration
• Parallel recognizers and “understanders”
• Time-stamped meaning fragments for each
stream
• Common framework for meaning
representation – typed feature structures
• Meaning fusion operation -- unification
• Process for determining a joint interpretation
(subject to semantic, and spatiotemporal constraints)
• Statistical ranking
• Flexible asynchronous architecture
• Must handle unimodal and multimodal input
From speech (one of many hyp’s)
“Evacuation route”
Color: green
Object:
Label:
Evacuation route
Line_obj
Create_line
Color: green
Location: Line [ ]
Object:
Label:
Evacuation route
Line_obj
From sketch
Location:
ISA
Location:
Coordlist:
[(95302,94360),
(95305,94365)],
Line
Create_line
…]
Line
command
[location:
command
point
[Xcoord: 95305, Ycoord: 94365 ]]
Coordlist:
[(95302,94360),
(95305,94365)],
…]
Mutual
Disambiguation
• Each input mode
provides a set of
scored recognition
hypotheses
speech gesture object multimodal
s1
g1
o1
mm1
s2
g2
o2
mm2
s3
g3
o3
mm3
g4
mm4
• MD derives the best joint interpretation by unification of
meaning representation fragments
• PMM = αPS + βPG + C; learn α, β, and C over a multimodal corpus
• MD stabilizes system performance in challenging environments
Benefits of mutual disambiguation
Application
RER
Reference
Non-native speakers
and moderate mobility
19-41%
multimodal cmds
Oviatt ‘99
Exerted users
35%
multimodal cmds
Kumar et al., ICMI,
2004
Multimodal 3D AR/VR
environments
67%
multimodal cmds
Kaiser et al., 2003
New vocabulary
speech and
handwriting
Audiovisual speech
recog in noisy
environments
66% phoneme
16% HW
35-50%
Words
Kaiser et al., 2004,
and Kaiser PhD thesis
Potamianos, Neti et
al. 2003
Efficiency Benefits
Time (secs.) to create
entity or repair errors
Efficiency Comparison
CPOF
50
40
30
20
10
0
MM
GUI
Units
Lines
& Areas
Control
measures
Repair
errors
MM 16x faster
(NIS)
Demonstration
CMU -speech
MIT – body
tracking
OHSU –
multimodal
fusion
(speech +
writing/sketch,
3D gesture)
Stanford
(NLP, dialogue)
Tangible Multimodal Systems for SafetyCritical Applications
What’s Missing?
A Division
Command
Post during
an exercise
McGee et al., CHI ‘02;
Cohen & McGee,
CACM’04
What they use
Many work practices rely on paper
ATC -- Mackay ‘98
ICU -- Gorman et al., 2000
Why do they
use paper?
• Already know the
interface
• Poor computer
interfaces
• Fail-safe; robust to
power outages
• High resolution
• Large/small scale
• Cheap
• Lightweight
• Portable
• Collaboration
Clinical Data Entry
“Perhaps the single greatest challenge that has
consistently confronted every clinical system
developer is to engage clinicians in direct data
entry” (IOM, 1997, p. 125)
“To make it simple for the practitioner to interact
with the record, data entry must be almost as
easy as writing.” (IOM. 1997, p. 88)
Multimodal Interaction with Paper
(NIS)
Based on Anoto technology
Benefits
• Most people (incl. kids, seniors) know
how to use the pen
• Portability (works over cell phone)
• Ubiquity – paper is everywhere
• Collaborative – multiple simult. pens
• Next – use for note-taking, alone or in
meetings; fuse with ongoing speech
• Many new applications – e.g.,
architecture, engineering, education,
field data capture
Elementary Science Education
Sharon Oviatt
Quiet Interfaces that Help People Think
Sharon Oviatt
[email protected]
http://www.cse.ogi.edu/CHCC/