Title :
Lip animation of a personalized facial model from auditory speech
Author :
Melenchón, Javier ; Iriondo, Ignasi ; Socoró, J. Claudi ; Matinez, E. ; Meler, Lourdes
Author_Institution :
La Salle Sch. of Eng., Ramon Llull Univ., Barcelona, Spain
Abstract :
This paper proposes a new method for lip animation of personalized facial model from auditory speech. It is based on Bayesian estimation and person specific appearance models (PSFAM). Initially, a video of a speaking person is recorded from which the visual and acoustic features of the speaker and their relationship will be learnt. First, the visual information of the speaker is stored in a color PSFAM by means of a registration algorithm. Second, the auditory features are extracted from the waveform attached to the recorded video sequence. Third, the relationship between the learnt PSFAM and the auditory features of the speaker is represented by Bayesian estimators. Finally, subjective perceptual tests are reported in order to measure the intelligibility of the preliminary results synthesizing isolated words.
Keywords :
Bayes methods; computer animation; feature extraction; image sequences; least mean squares methods; speech synthesis; vector quantisation; video recording; video signal processing; Bayesian estimation; acoustic features; auditory speech; color PSFAM; lip animation; person specific appearance models; personalized facial model; recorded video sequence; registration algorithm; Artificial neural networks; Bayesian methods; Data mining; Facial animation; Feature extraction; Head; Hidden Markov models; Loudspeakers; Speech synthesis; Testing;
Conference_Titel :
Signal Processing and Information Technology, 2003. ISSPIT 2003. Proceedings of the 3rd IEEE International Symposium on
Print_ISBN :
0-7803-8292-7
DOI :
10.1109/ISSPIT.2003.1341108