Title :
Text-to-visual speech synthesis based on parameter generation from HMM
Author :
Masuko, Takashi ; Kobayashi, Takao ; Tamura, Masatsune ; Masubuchi, Jun ; Tokuda, Keiichi
Author_Institution :
Precision & Intell. Lab., Tokyo Inst. of Technol., Japan
Abstract :
This paper presents a new technique for synthesizing visual speech from arbitrarily given text. The technique is based on an algorithm for parameter generation from HMM with dynamic features, which has been successfully applied to text-to-speech synthesis. In the training phase, syllable HMMs are trained with visual speech parameter sequences that represent lip movements. In the synthesis phase, a sentence HMM is constructed by concatenating syllable HMMs corresponding to the phonetic transcription for the input text. Then an optimum visual speech parameter sequence is generated from the sentence HMM in an ML sense. The proposed technique can generate synchronized lip movements with speech in a unified framework. Furthermore, coarticulation is implicitly incorporated into the generated mouth shapes. As a result, synthetic lip motion becomes smooth and realistic
Keywords :
computer animation; feature extraction; hidden Markov models; speech synthesis; synchronisation; 2D lip contour animation; ML; coarticulation; dynamic features; generated mouth shapes; input text; lip movements; parameter generation algorithm; phonetic transcription; sentence HMM; syllable HMM; synchronized lip movements; synthesis phase; synthetic lip motion; text-to-visual speech synthesis; training phase; visual speech parameter sequences; Audio databases; Hidden Markov models; Laboratories; Mouth; Shape; Spatial databases; Speech synthesis; Statistics; Videoconference; Visual databases;
Conference_Titel :
Acoustics, Speech and Signal Processing, 1998. Proceedings of the 1998 IEEE International Conference on
Conference_Location :
Seattle, WA
Print_ISBN :
0-7803-4428-6
DOI :
10.1109/ICASSP.1998.679698