3-D facial pose and gaze point estimation using a robust real-time tracking paradigm
被引:52
作者:
Heinzmann, J
论文数: 0引用数: 0
h-index: 0
机构:
Australian Natl Univ, Dept Syst Engn, Res Sch Informat Sci & Engn, Canberra, ACT 0200, AustraliaAustralian Natl Univ, Dept Syst Engn, Res Sch Informat Sci & Engn, Canberra, ACT 0200, Australia
Heinzmann, J
[1
]
Zelinsky, A
论文数: 0引用数: 0
h-index: 0
机构:
Australian Natl Univ, Dept Syst Engn, Res Sch Informat Sci & Engn, Canberra, ACT 0200, AustraliaAustralian Natl Univ, Dept Syst Engn, Res Sch Informat Sci & Engn, Canberra, ACT 0200, Australia
Zelinsky, A
[1
]
机构:
[1] Australian Natl Univ, Dept Syst Engn, Res Sch Informat Sci & Engn, Canberra, ACT 0200, Australia
来源:
AUTOMATIC FACE AND GESTURE RECOGNITION - THIRD IEEE INTERNATIONAL CONFERENCE PROCEEDINGS
|
1998年
关键词:
D O I:
10.1109/AFGR.1998.670939
中图分类号:
TP18 [人工智能理论];
学科分类号:
081104 ;
0812 ;
0835 ;
1405 ;
摘要:
Facial pose and gaze point are fundamental to any visually directed human-machine interface. In this paper we propose a system capable of tracking a face and estimating the 3-D pose and the gaze point all in a real-time video stream of the head. This is done by using a 3-D model together with multiple triplet triangulation of feature positions assuming an affine projection. Using feature-based tracking the calculation of a 3-D eye gaze direction vector is possible even with head rotation and using a monocular camera. The system is also able to automatically initialise the feature tracking and to recover from total tracking failures which can occur when a person becomes occluded or temporarily leaves the image.