Title :
3D facial expression editing based on the dynamic graph model
Author :
Pei, Yuru ; Zha, Hongbin
Author_Institution :
Key Lab. of Machine Perception (MOE), Peking Univ., Beijing, China
fDate :
June 28 2009-July 3 2009
Abstract :
To model a detailed 3D expressive face based on the limited user constraints is a challenge work. In this paper, we present the facial expression editing technique based on a dynamic graph model. The probabilistic relations between facial expressions and the complex combination of local facial features, as well as the temporal behaviors of facial expressions are represented by the hierarchical dynamic Bayesian network. Given limited user-constraints on the sparse feature mesh, the system can infer the basis expression probabilities, which are used to locate the corresponding expressive mesh in the shape space spanned by the basis models. The experiments demonstrate the 3D dense facial meshes corresponding to the user-constraints can be synthesized effectively.
Keywords :
belief networks; graph theory; image processing; 3D facial expression editing; basis expression probabilities; dynamic graph model; hierarchical dynamic Bayesian network; local facial features; sparse feature mesh; Bayesian methods; Context modeling; Facial features; Facial muscles; Hidden Markov models; Humans; Laboratories; Network synthesis; Shape; Speech synthesis; Bayesian network; Expression editing;
Conference_Titel :
Multimedia and Expo, 2009. ICME 2009. IEEE International Conference on
Conference_Location :
New York, NY
Print_ISBN :
978-1-4244-4290-4
Electronic_ISBN :
1945-7871
DOI :
10.1109/ICME.2009.5202754