Title :
A Video, Text, and Speech-Driven Realistic 3-D Virtual Head for Human–Machine Interface
Author :
Jun Yu ; Zeng-Fu Wang
Author_Institution :
Nat. Lab. of Speech & Language Inf. Process., Univ. of Sci. & Technol. of China, Hefei, China
Abstract :
A multiple inputs-driven realistic facial animation system based on 3-D virtual head for human-machine interface is proposed. The system can be driven independently by video, text, and speech, thus can interact with humans through diverse interfaces. The combination of parameterized model and muscular model is used to obtain a tradeoff between computational efficiency and high realism of 3-D facial animation. The online appearance model is used to track 3-D facial motion from video in the framework of particle filtering, and multiple measurements, i.e., pixel color value of input image and Gabor wavelet coefficient of illumination ratio image, are infused to reduce the influence of lighting and person dependence for the construction of online appearance model. The tri-phone model is used to reduce the computational consumption of visual co-articulation in speech synchronized viseme synthesis without sacrificing any performance. The objective and subjective experiments show that the system is suitable for human-machine interaction.
Keywords :
computer animation; human computer interaction; image colour analysis; man-machine systems; particle filtering (numerical methods); speech synthesis; text detection; video signal processing; wavelet transforms; 3-D facial motion; Gabor wavelet coefficient; computational consumption; human-machine interface; illumination ratio image; multiple input-driven realistic facial animation system; online appearance model; particle filtering; pixel color value; speech synchronized viseme synthesis; speech-driven realistic 3-D virtual head; text-driven realistic 3-D virtual head; triphone model; video-driven realistic 3-D virtual head; visual coarticulation; Adaptation models; Computational modeling; Face; Facial animation; Magnetic heads; Solid modeling; Speech; Facial animation; facial motion tracking; human–machine interface; human???machine interface; virtual head;
Journal_Title :
Cybernetics, IEEE Transactions on
DOI :
10.1109/TCYB.2014.2341737