Title :
Multimodal speech animation from electromagnetic articulography data
Author :
Gibert, Guillaume ; Attina, V. ; Tiede, M. ; Bundgaard-Nielsen, R. ; Kroos, C. ; Kasisopa, B. ; Vatikiotis-Bateson, E. ; Best, C.T.
Author_Institution :
U846, Inserm, Bron, France
Abstract :
Virtual humans have become part of our everyday life (movies, internet, and computer games). Even though they are more and more realistic, their speech capabilities are, most of the time, limited and not coherent and/or not synchronous with the corresponding acoustic signal. We describe a method to convert a virtual human avatar (animated through key frames and interpolation) into a more naturalistic talking head. Speech-capabilities were added to the avatar using real speech production data. Electromagnetic articulography (EMA) data provided lip, jaw and tongue trajectories of a speaker involved in face to face communication. An articulatory model driving jaw, lip and tongue movements was built. Constraining the key frame values, a corresponding high definition tongue articulatory model was developed. The resulting avatar was able to produce visible and partly occluded facial speech movements coherent and synchronous with the acoustic signal.
Keywords :
avatars; computer animation; speech processing; EMA; acoustic signal; electromagnetic articulography data; face to face communication; jaw trajectories; multimodal speech animation; speech capabilities; tongue trajectories; virtual human avatar; Animation; Avatars; Face; Lips; Principal component analysis; Speech; Tongue; Augmented speech; ElectroMagnetic Articulography (EMA); Speech production; Talking head;
Conference_Titel :
Signal Processing Conference (EUSIPCO), 2012 Proceedings of the 20th European
Conference_Location :
Bucharest
Print_ISBN :
978-1-4673-1068-0