DocumentCode :
1349947
Title :
Toward multimodal human-computer interface
Author :
Sharma, Rejeev ; Pavlovic, Vladimir I. ; Huang, Thomas S.
Author_Institution :
Dept. of Comput. Sci. & Eng., Pennsylvania State Univ., University Park, PA, USA
Volume :
86
Issue :
5
fYear :
1998
fDate :
5/1/1998 12:00:00 AM
Firstpage :
853
Lastpage :
869
Abstract :
Recent advances in various signal processing technologies, coupled with an explosion in the available computing power, have given rise to a number of novel human-computer interaction (HCI) modalities: speech, vision-based gesture recognition, eye tracking, electroencephalograph, etc. Successful embodiment of these modalities into an interface has the potential of easing the HCI bottleneck that has become noticeable with the advances in computing and communication. It has also become increasingly evident that the difficulties encountered in the analysis and interpretation of individual sensing modalities may be overcome by integrating them into a multimodal human-computer interface. We examine several promising directions toward achieving multimodal HCI. We consider some of the emerging novel input modalities for HCI and the fundamental issues in integrating them at various levels, from early signal level to intermediate feature level to late decision level. We discuss the different computational approaches that may be applied at the different levels of modality integration. We also briefly review several demonstrated multimodal HCI systems and applications. Despite all the recent developments, it is clear that further research is needed for interpreting and fitting multiple sensing modalities in the context of HCI. This research can benefit from many disparate fields of study that increase our understanding of the different human communication modalities and their potential role in HCI
Keywords :
human factors; image recognition; speech recognition; user interfaces; HCI bottleneck; computational approaches; electroencephalograph; eye tracking; gesture recognition; input modalities; modality integration; multimodal human-computer interface; research; signal processing; speech; Artificial neural networks; Automatic speech recognition; Computer displays; Computer interfaces; Electroencephalography; Hidden Markov models; Human computer interaction; Maximum likelihood estimation; Personal digital assistants; Virtual reality;
fLanguage :
English
Journal_Title :
Proceedings of the IEEE
Publisher :
ieee
ISSN :
0018-9219
Type :
jour
DOI :
10.1109/5.664275
Filename :
664275
Link To Document :
بازگشت