DocumentCode
15014
Title
A Video, Text, and Speech-Driven Realistic 3-D Virtual Head for Human–Machine Interface
Author
Jun Yu ; Zeng-Fu Wang
Author_Institution
Nat. Lab. of Speech & Language Inf. Process., Univ. of Sci. & Technol. of China, Hefei, China
Volume
45
Issue
5
fYear
2015
fDate
May-15
Firstpage
977
Lastpage
988
Abstract
A multiple inputs-driven realistic facial animation system based on 3-D virtual head for human-machine interface is proposed. The system can be driven independently by video, text, and speech, thus can interact with humans through diverse interfaces. The combination of parameterized model and muscular model is used to obtain a tradeoff between computational efficiency and high realism of 3-D facial animation. The online appearance model is used to track 3-D facial motion from video in the framework of particle filtering, and multiple measurements, i.e., pixel color value of input image and Gabor wavelet coefficient of illumination ratio image, are infused to reduce the influence of lighting and person dependence for the construction of online appearance model. The tri-phone model is used to reduce the computational consumption of visual co-articulation in speech synchronized viseme synthesis without sacrificing any performance. The objective and subjective experiments show that the system is suitable for human-machine interaction.
Keywords
computer animation; human computer interaction; image colour analysis; man-machine systems; particle filtering (numerical methods); speech synthesis; text detection; video signal processing; wavelet transforms; 3-D facial motion; Gabor wavelet coefficient; computational consumption; human-machine interface; illumination ratio image; multiple input-driven realistic facial animation system; online appearance model; particle filtering; pixel color value; speech synchronized viseme synthesis; speech-driven realistic 3-D virtual head; text-driven realistic 3-D virtual head; triphone model; video-driven realistic 3-D virtual head; visual coarticulation; Adaptation models; Computational modeling; Face; Facial animation; Magnetic heads; Solid modeling; Speech; Facial animation; facial motion tracking; human–machine interface; human???machine interface; virtual head;
fLanguage
English
Journal_Title
Cybernetics, IEEE Transactions on
Publisher
ieee
ISSN
2168-2267
Type
jour
DOI
10.1109/TCYB.2014.2341737
Filename
6872579
Link To Document