Skip to main content

1999 | OriginalPaper | Buchkapitel

Multimodal Interfaces for Multimedia Information Agents

verfasst von : Alex Waibel, Bernhard Suhm, Minh Tue Vo, Jie Yang

Erschienen in: Computational Models of Speech Pattern Processing

Verlag: Springer Berlin Heidelberg

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

When humans communicate they take advantage of a rich spectrum of cues. Some are verbal and acoustic. Some are non-verbal and non-acoustic. Signal processing technology has devoted much attention to the recognition of speech, as a single human communication signal. Most other complementary communication cues, however, remain unexplored and unused in human-computer interaction. In this paper we show that the addition of non-acoustic or non-verbal cues can significantly enhance robustness, flexibility, naturalness and performance of human-computer interaction. We demonstrate computer agents that use speech, gesture, handwriting, pointing, spelling jointly for more robust, natural and flexible human-computer interaction in the various tasks of an information worker: information creation, access, manipulation or dissemination.

Metadaten
Titel
Multimodal Interfaces for Multimedia Information Agents
verfasst von
Alex Waibel
Bernhard Suhm
Minh Tue Vo
Jie Yang
Copyright-Jahr
1999
Verlag
Springer Berlin Heidelberg
DOI
https://doi.org/10.1007/978-3-642-60087-6_35

Neuer Inhalt