DocumentCode
457096
Title
Speech Animation Using Coupled Hidden Markov Models
Author
Xie, Lei ; Liu, Zhi-Qiang
Author_Institution
Sch. of Creative Media, Hong Kong City Univ.
Volume
1
fYear
0
fDate
0-0 0
Firstpage
1128
Lastpage
1131
Abstract
We present a novel speech animation approach using coupled hidden Markov models (CHMMs). Different from the conventional HMMs that use a single state chain to model the audio-visual speech with tight inter-modal synchronization, we use the CHMMs to model the asynchrony, different discriminative abilities, and temporal coupling between the audio speech and the visual speech, which are important factors for animations looking natural. Based on the audio-visual CHMMs, visual animation parameters are predicted from audio through an EM-based audio to visual conversion algorithm. Experiments on the JEWEL AV database show that compared with the conventional HMMs, the CHMMs can output visual parameters that are much closer to the actual ones. Explicit modelling of audio-visual speech is promising in speech animation
Keywords
computer animation; hidden Markov models; speech processing; audio speech animation; audio-visual speech; coupled hidden Markov model; explicit modelling; intermodal synchronization; visual animation parameter; visual speech; Application software; Augmented virtuality; Face; Facial animation; Hidden Markov models; Multimedia databases; Speech processing; Speech synthesis; Visual databases; Viterbi algorithm;
fLanguage
English
Publisher
ieee
Conference_Titel
Pattern Recognition, 2006. ICPR 2006. 18th International Conference on
Conference_Location
Hong Kong
ISSN
1051-4651
Print_ISBN
0-7695-2521-0
Type
conf
DOI
10.1109/ICPR.2006.1074
Filename
1699088
Link To Document