共 59 条
[1]
Bentivoglio AR(1997)Analysis of blink rate patterns in normal subjects Movement Disorders 12 1028-1034
[2]
Bressman SB(2014)CREMA-D: Crowd-sourced emotional multimodal actors dataset IEEE Transactions on Affective Computing 5 377-390
[3]
Cassetta E(2005)Expressive speech-driven facial animation ACM TOG 24 1283-1302
[4]
Carretta D(2006)An audio-visual corpus for speech perception and automatic speech recognition The Journal of the Acoustical Society of America 120 2421-2424
[5]
Tonali P(2015)TCD-TIMIT: An audio-visual corpus of continuous speech IEEE Transactions on Multimedia 17 603-615
[6]
Albanese A(2017)Audio-driven facial animation by joint end-to-end learning of pose and emotion ACM TOG 36 1-12
[7]
Cao H(2009)A no-reference perceptual image sharpness metric based on a cumulative probability of blur detection International Workshop on Quality of Multimedia Experience (QoMEx) 20 87-91
[8]
Cooper DG(1990)Generation of mouthshapes for a synthetic talking head Proceedings of the Institute of Acoustics, Autumn Meeting 12 475-482
[9]
Keutmann MK(2017)Synthesizing Obama: Learning lip sync from audio output Obama video ACM TOG 36 1-13
[10]
Gur RC(2017)A deep learning approach for generalized speech animation ACM TOG 36 1-13