DocumentCode :
179227
Title :
Evaluation of HMM-based visual laughter synthesis
Author :
Cakmak, Huseyin ; Urbain, Jerome ; Tilmanne, Joelle ; Dutoit, Thierry
Author_Institution :
TCTS Lab., Univ. of Mons, Mons, Belgium
fYear :
2014
fDate :
4-9 May 2014
Firstpage :
4578
Lastpage :
4582
Abstract :
In this paper we apply speaker-dependent training of Hidden Markov Models (HMMs) to audio and visual laughter synthesis separately. The two modalities are synthesized with a forced durations approach and are then combined together to render audio-visual laughter on a 3D avatar. This paper focuses on visual synthesis of laughter and its perceptive evaluation when combined with synthesized audio laughter. Previous work on audio and visual synthesis has been successfully applied to speech. The extrapolation to audio laughter synthesis has already been done. This paper shows that it is possible to extrapolate to visual laughter synthesis as well.
Keywords :
audio-visual systems; avatars; extrapolation; hidden Markov models; speech synthesis; 3D avatar; HMM based visual laughter synthesis evaluation; audio-visual laughter synthesis; extrapolation; hidden Markov model; speaker-dependent training; Databases; Face; Hidden Markov models; Pipelines; Speech; Videos; Visualization; Audio; HMM; laughter; synthesis; visual;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2014 IEEE International Conference on
Conference_Location :
Florence
Type :
conf
DOI :
10.1109/ICASSP.2014.6854469
Filename :
6854469
Link To Document :
بازگشت