DocumentCode
3741867
Title
Lip movement and speech synchronization detection based on multimodal shift-invariant dictionary
Author
Zhengyu Zhu; Xiaohui Feng; Jichen Yang
Author_Institution
South China University of Technology, Guangzhou 510641, China
fYear
2015
Firstpage
768
Lastpage
772
Abstract
In order to solve the issue of ignoring the successive and dynamic lip motion information in traditional audio-visual speech synchrony analysis models, a novel method based on shift-invariant learned dictionary is presented. In this method, sparse representation with shift-invariant dictionary is introduced to analyze the bimodal structure of articulation. The learned dictionary is obtained based on the audio-visual coherence dictionary learning algorithm, and the dynamic correlation between voice and lip motion of diverse syllable or word is represented as a pattern by this audio-visual coherence atom. According to these utterance patterns, an original audio-visual synchronization score measuring scheme is proposed. The results of the experiment on four different asynchronous data show the good performance of the method.
Keywords
"Synchronization","Chlorine","Discrete cosine transforms","Coal"
Publisher
ieee
Conference_Titel
Communication Technology (ICCT), 2015 IEEE 16th International Conference on
Print_ISBN
978-1-4673-7004-2
Type
conf
DOI
10.1109/ICCT.2015.7399944
Filename
7399944
Link To Document