• DocumentCode
    3741867
  • Title

    Lip movement and speech synchronization detection based on multimodal shift-invariant dictionary

  • Author

    Zhengyu Zhu; Xiaohui Feng; Jichen Yang

  • Author_Institution
    South China University of Technology, Guangzhou 510641, China
  • fYear
    2015
  • Firstpage
    768
  • Lastpage
    772
  • Abstract
    In order to solve the issue of ignoring the successive and dynamic lip motion information in traditional audio-visual speech synchrony analysis models, a novel method based on shift-invariant learned dictionary is presented. In this method, sparse representation with shift-invariant dictionary is introduced to analyze the bimodal structure of articulation. The learned dictionary is obtained based on the audio-visual coherence dictionary learning algorithm, and the dynamic correlation between voice and lip motion of diverse syllable or word is represented as a pattern by this audio-visual coherence atom. According to these utterance patterns, an original audio-visual synchronization score measuring scheme is proposed. The results of the experiment on four different asynchronous data show the good performance of the method.
  • Keywords
    "Synchronization","Chlorine","Discrete cosine transforms","Coal"
  • Publisher
    ieee
  • Conference_Titel
    Communication Technology (ICCT), 2015 IEEE 16th International Conference on
  • Print_ISBN
    978-1-4673-7004-2
  • Type

    conf

  • DOI
    10.1109/ICCT.2015.7399944
  • Filename
    7399944