SAVI: an actively controlled teleconferencing system

被引:8
作者
Herpers, R
Derpanis, K
MacLean, WJ
Verghese, G
Jenkin, M
Milios, E
Jepson, A
Tsotsos, JK
机构
[1] Univ Appl Sci St Augustin, Fachhsch Bonn Rhein Sieg, Dept Appl Comp Sci, D-53757 St Augustin, Germany
[2] York Univ, Dept Comp Sci, York Ctr Vis Res, N York, ON M3J 1P3, Canada
[3] Univ Toronto, Dept Elect & Comp Engn, Toronto, ON M5S 3G4, Canada
[4] Univ Toronto, Dept Comp Sci, Toronto, ON M5S 3G4, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
active vision interface; teleconferencing system; face and hand gesture recognition; Computer Vision System;
D O I
10.1016/S0262-8856(00)00107-4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A Stereo Active Vision Interface (SAVI) is introduced which detects frontal faces in real world environments and performs particular active control tasks dependent on hand gestures given by the person the system attends to. The SAVI system is thought of as a smart user interface for teleconferencing, telemedicine, and distance learning applications. To reduce the search space in the visual scene the processing is started with the detection of connected skin colour regions applying a new radial scanline algorithm. Subsequently, in the most salient skin colour region facial features are searched for while the skin colour blob is actively kept in the centre of the visual field of the camera system. After a successful evaluation of the facial features the associated person is able to give control commands to the system. For this contribution only visual control commands are investigated but there is no limitation for voice or any other commands. These control commands can either effect the observing system itself or any other active or robotic system wired to the principle observing system via TCP/IP sockets. The system is designed as a perception-action-cycle (PAC), processing sensory data of different kinds and qualities. Both the vision module and the head motion control module work at frame rate on a PC platform. Hence, the system is able to react instantaneously to changing conditions in the visual scene. (C) 2001 Elsevier Science B.V. All rights reserved.
引用
收藏
页码:793 / 804
页数:12
相关论文
共 23 条
  • [11] Jennings C., 1999, Proceedings International Workshop on Recognition, Analysis, and Tracking of Faces and Gestures in Real-Time Systems. In Conjunction with ICCV'99 (Cat. No.PR00378), P152, DOI 10.1109/RATFG.1999.799238
  • [12] Kender J., 1995, P INT WORKSH AUT FAC, P184
  • [13] Toward the use of gesture in traditional user interfaces
    Kjeldsen, R
    Kender, J
    [J]. PROCEEDINGS OF THE SECOND INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION, 1996, : 151 - 156
  • [14] Finding skin in color images
    Kjeldsen, R
    Kender, J
    [J]. PROCEEDINGS OF THE SECOND INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION, 1996, : 312 - 317
  • [15] KLETTE R, 1996, HDB IMAGE PROCESSING
  • [16] LUCENTE M, 1998, INT ENV S
  • [17] Milios E., 1993, International Journal of Pattern Recognition and Artificial Intelligence, V7, P51, DOI 10.1142/S0218001493000042
  • [18] Tracking and segmenting people in varying lighting conditions using colour
    Raja, Y
    McKenna, SJ
    Gong, SG
    [J]. AUTOMATIC FACE AND GESTURE RECOGNITION - THIRD IEEE INTERNATIONAL CONFERENCE PROCEEDINGS, 1998, : 228 - 233
  • [19] PLAYBOT - A visually-guided robot for physically disabled children
    Tsotsos, JK
    Verghese, G
    Dickinson, S
    Jenkin, M
    Jepson, A
    Milios, E
    Nuflo, F
    Stevenson, S
    Black, M
    Metaxas, D
    Culhane, S
    Ye, Y
    Mann, R
    [J]. IMAGE AND VISION COMPUTING, 1998, 16 (04) : 275 - 292
  • [20] EIGENFACES FOR RECOGNITION
    TURK, M
    PENTLAND, A
    [J]. JOURNAL OF COGNITIVE NEUROSCIENCE, 1991, 3 (01) : 71 - 86