Modelling Spatio-Temporal Saliency to Predict Gaze Direction for Short Videos

被引:155
作者
Marat, Sophie [1 ]
Phuoc, Tien Ho [1 ]
Granjon, Lionel [1 ]
Guyader, Nathalie [1 ]
Pellerin, Denis [1 ]
Guerin-Dugue, Anne [1 ]
机构
[1] Dept Images Signal, GIPSA Lab, F-38402 Grenoble, France
关键词
Saliency; Spatio-temporal model; Gaze prediction; Video viewing; VISUAL-ATTENTION; INTEGRATION; ALLOCATION; SELECTION;
D O I
10.1007/s11263-009-0215-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents a spatio-temporal saliency model that predicts eye movement during video free viewing. This model is inspired by the biology of the first steps of the human visual system. The model extracts two signals from video stream corresponding to the two main outputs of the retina: parvocellular and magnocellular. Then, both signals are split into elementary feature maps by cortical-like filters. These feature maps are used to form two saliency maps: a static and a dynamic one. These maps are then fused into a spatio-temporal saliency map. The model is evaluated by comparing the salient areas of each frame predicted by the spatio-temporal saliency map to the eye positions of different subjects during a free video viewing experiment with a large database (17000 frames). In parallel, the static and the dynamic pathways are analyzed to understand what is more or less salient and for what type of videos our model is a good or a poor predictor of eye movement.
引用
收藏
页码:231 / 243
页数:13
相关论文
共 35 条