Title :
Internet communication using real-time facial expression analysis and synthesis
Author :
Chandrasiri, Naiwala P. ; Naemura, Takeshi ; Ishizuka, Mitsuru ; Harashima, Hiroshi ; Barakonyi, István
Author_Institution :
Tokyo Univ., Japan
Abstract :
In this paper, the authors have developed a system that animates 3D facial agents based on real-time facial expression analysis techniques and research on synthesizing facial expressions and text-to-speech capabilities. This system combines visual, auditory, and primary interfaces to communicate one coherent multimodal chat experience. Users can represent themselves using agents they select from a group that we have predefined. When a user shows a particular expression while typing a text, the 3D agent at the receiving end speaks the message aloud while it replays the recognized facial expression sequences and also augments the synthesized voice with appropriate emotional content. Because the visual data exchange is based on the MPEG-4 high-level Facial Animation Parameter for facial expressions (FAP 2), rather than real-time video, the method requires very low bandwidth.
Keywords :
Internet; electronic messaging; face recognition; real-time systems; software agents; Internet communication; MPEG-4 high-level Facial Animation Parameter; auditory interfaces; emotional content; facial agents; facial animation; facial expression analysis; facial expression sequences; facial expression synthesis; facial recognition; multimodal chat; primary interfaces; real-time analysis; real-time synthesis; text-to-speech capabilities; visual data exchanging; visual interfaces; voice synthesis; Emotion recognition; Face recognition; Facial animation; Financial advantage program; Internet; MPEG 4 Standard; Real time systems; Speech recognition; Speech synthesis; Text recognition;
Journal_Title :
MultiMedia, IEEE
DOI :
10.1109/MMUL.2004.10