Title of article :
Transferring of Speech Movements from Video to 3D Face Space
Author/Authors :
Yuru Pei، نويسنده , , Hongbin Zha، نويسنده ,
Issue Information :
روزنامه با شماره پیاپی سال 2007
Abstract :
We present a novel method for transferring speech animation recorded in low quality videos to high resolution 3D face
models. The basic idea is to synthesize the animated faces by an interpolation based on a small set of 3D key face shapes which span
a 3D face space. The 3D key shapes are extracted by an unsupervised learning process in 2D video space to form a set of 2D visemes
which are then mapped to the 3D face space. The learning process consists of two main phases: 1) Isomap-based nonlinear
dimensionality reduction to embed the video speech movements into a low-dimensional manifold and 2) K-means clustering in the lowdimensional
space to extract 2D key viseme frames. Our main contribution is that we use the Isomap-based learning method to extract
intrinsic geometry of the speech video space and thus to make it possible to define the 3D key viseme shapes. To do so, we need only
to capture a limited number of 3D key face models by using a general 3D scanner. Moreover, we also develop a skull movement
recovery method based on simple anatomical structures to enhance 3D realism in local mouth movements. Experimental results show
that our method can achieve realistic 3D animation effects with a small number of 3D key face models.
Keywords :
facial animation , visual speech synthesis , performance-driven animation , speech synchronization , machine learning.
Journal title :
IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS
Journal title :
IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS