Toward multimodal human-computer interface

被引:165
作者
Sharma, R [1 ]
Pavlovic, VI
Huang, TS
机构
[1] Penn State Univ, Dept Comp Sci & Engn, University Pk, PA 16802 USA
[2] Univ Illinois, Beckman Inst, Urbana, IL 61801 USA
[3] Univ Illinois, Dept Elect & Comp Engn, Urbana, IL 61801 USA
基金
美国国家科学基金会;
关键词
human-computer interface; multimodality; sensor fusion;
D O I
10.1109/5.664275
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Recent advances in various signal-processing technologies, coupled with an explosion in the available computing power, have given rise to a number of novel human-computer interaction (HCI) modalities-speech, vision-based gesture recognition, eye tracking, electroencephalograph, etc. Successful embodiment of these modalities into an interface has the potential of easing the HCI bottleneck that has become noticeable with the advances in computing and communication. It has also become increasingly evident that the difficulties encountered in the analysis and interpretation of individual sensing modalities may be overcome by integrating them into a multimodal human-computer interface. In this paper, we examine several promising directions toward achieving multimodal HCI. We consider some of the emerging novel input modalities for HCI and the fundamental issues in integrating them at various levels-from early "signal" level to intermediate "feature" level to late "decision" level. We discuss the different computational approaches that may be applied at the different levels of modality integration. We also briefly review several demonstrated multimodal HCI system and applications. Despite all the recent developments, it is clear that further research is needed for interpreting and fusing multiple sensing modalities in the context of HCI. This research can benefit from many disparate fields of study that increase our understanding of the different human communication modalities and their potential role in HCI.
引用
收藏
页码:853 / 869
页数:17
相关论文
共 93 条
[1]  
ADAM JA, 1993, IEEE SPECTRUM, V30, P22, DOI 10.1109/6.237580
[2]  
ADJOUDANI A, 1995, P EUR C MADR SPAIN, V2, P1563
[3]  
[Anonymous], SEMANTIC NETWORKS AR
[4]  
[Anonymous], 1997, Software Agents
[5]  
[Anonymous], PERCEPTION ESSAYS HO
[6]  
[Anonymous], P INT WORKSH FAC GES
[7]  
[Anonymous], EVIDENCE THEORY ITS
[8]  
[Anonymous], SOFTWARE AGENTS
[9]   TRACKING REQUIREMENTS FOR AUGMENTED REALITY [J].
AZUMA, R .
COMMUNICATIONS OF THE ACM, 1993, 36 (07) :50-51
[10]   CHARADE - REMOTE-CONTROL OF OBJECTS USING FREE-HAND GESTURES [J].
BAUDEL, T ;
BEAUDOUINLAFON, M .
COMMUNICATIONS OF THE ACM, 1993, 36 (07) :28-35