• DocumentCode
    15014
  • Title

    A Video, Text, and Speech-Driven Realistic 3-D Virtual Head for Human–Machine Interface

  • Author

    Jun Yu ; Zeng-Fu Wang

  • Author_Institution
    Nat. Lab. of Speech & Language Inf. Process., Univ. of Sci. & Technol. of China, Hefei, China
  • Volume
    45
  • Issue
    5
  • fYear
    2015
  • fDate
    May-15
  • Firstpage
    977
  • Lastpage
    988
  • Abstract
    A multiple inputs-driven realistic facial animation system based on 3-D virtual head for human-machine interface is proposed. The system can be driven independently by video, text, and speech, thus can interact with humans through diverse interfaces. The combination of parameterized model and muscular model is used to obtain a tradeoff between computational efficiency and high realism of 3-D facial animation. The online appearance model is used to track 3-D facial motion from video in the framework of particle filtering, and multiple measurements, i.e., pixel color value of input image and Gabor wavelet coefficient of illumination ratio image, are infused to reduce the influence of lighting and person dependence for the construction of online appearance model. The tri-phone model is used to reduce the computational consumption of visual co-articulation in speech synchronized viseme synthesis without sacrificing any performance. The objective and subjective experiments show that the system is suitable for human-machine interaction.
  • Keywords
    computer animation; human computer interaction; image colour analysis; man-machine systems; particle filtering (numerical methods); speech synthesis; text detection; video signal processing; wavelet transforms; 3-D facial motion; Gabor wavelet coefficient; computational consumption; human-machine interface; illumination ratio image; multiple input-driven realistic facial animation system; online appearance model; particle filtering; pixel color value; speech synchronized viseme synthesis; speech-driven realistic 3-D virtual head; text-driven realistic 3-D virtual head; triphone model; video-driven realistic 3-D virtual head; visual coarticulation; Adaptation models; Computational modeling; Face; Facial animation; Magnetic heads; Solid modeling; Speech; Facial animation; facial motion tracking; human–machine interface; human???machine interface; virtual head;
  • fLanguage
    English
  • Journal_Title
    Cybernetics, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    2168-2267
  • Type

    jour

  • DOI
    10.1109/TCYB.2014.2341737
  • Filename
    6872579