Accurate, real-time, unadorned lip tracking

被引:27
作者
Kaucic, R [1 ]
Blake, A [1 ]
机构
[1] Univ Oxford, Dept Engn Sci, Oxford OX1 3PJ, England
来源
SIXTH INTERNATIONAL CONFERENCE ON COMPUTER VISION | 1998年
关键词
D O I
10.1109/ICCV.1998.710745
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Human speech is inherently multi-modal, consisting of both audio and visual components. Recently researchers have shown that the incorporation of information about the position of the lips into acoustic speech recognisers enables robust recognition of noisy speech. In the case of Hidden Markov Model-recognition, we show that this happens because the visual signal stabilises the alignment of states. A is also shown that unadorned lips, both the inner and outer contours, can be robustly tracked in real time on general-purpose workstations. To accomplish this, efficient algorithms are employed which contain three key components: shape models, motion models, and focused colo clr feature detectors - all of which are learnt from examples.
引用
收藏
页码:370 / 375
页数:6
相关论文
empty
未找到相关数据