DocumentCode :
61628
Title :
A Practical Model for Live Speech-Driven Lip-Sync
Author :
Li Wei ; Zhigang Deng
Author_Institution :
Dept. of Comput. Sci., Univ. of Houston, Houston, TX, USA
Volume :
35
Issue :
2
fYear :
2015
fDate :
Mar.-Apr. 2015
Firstpage :
70
Lastpage :
78
Abstract :
This article introduces a simple, efficient, yet practical phoneme-based approach to generate realistic speech animation in real time based on live speech input. Specifically, the authors first decompose lower-face movements into low-dimensional principal component spaces. Then, in each of the retained principal component spaces, they select the AnimPho with the highest priority value and the minimum smoothness energy. Finally, they apply motion blending and interpolation techniques to compute final animation frames for the currently inputted phoneme. Through many experiments and comparisons, the authors demonstrate the realism of synthesized speech animation by their approach as well as its real-time efficiency on an off-the-shelf computer.
Keywords :
computer animation; interpolation; speech processing; AnimPho; animation frames; interpolation techniques; live speech input; live speech-driven lip-sync; low-dimensional principal component spaces; lower-face movements; minimum smoothness energy; motion blending; phoneme-based approach; priority value; realistic speech animation; synthesized speech animation; Animation; Interpolation; Motion segmentation; Real-time systems; Speech processing; Speech recognition; computer graphics; facial animation; live speech driven; speech animation; talking avatars; virtual humans;
fLanguage :
English
Journal_Title :
Computer Graphics and Applications, IEEE
Publisher :
ieee
ISSN :
0272-1716
Type :
jour
DOI :
10.1109/MCG.2014.105
Filename :
6894485
Link To Document :
بازگشت