• DocumentCode
    2935240
  • Title

    3D facial expression editing based on the dynamic graph model

  • Author

    Pei, Yuru ; Zha, Hongbin

  • Author_Institution
    Key Lab. of Machine Perception (MOE), Peking Univ., Beijing, China
  • fYear
    2009
  • fDate
    June 28 2009-July 3 2009
  • Firstpage
    1354
  • Lastpage
    1357
  • Abstract
    To model a detailed 3D expressive face based on the limited user constraints is a challenge work. In this paper, we present the facial expression editing technique based on a dynamic graph model. The probabilistic relations between facial expressions and the complex combination of local facial features, as well as the temporal behaviors of facial expressions are represented by the hierarchical dynamic Bayesian network. Given limited user-constraints on the sparse feature mesh, the system can infer the basis expression probabilities, which are used to locate the corresponding expressive mesh in the shape space spanned by the basis models. The experiments demonstrate the 3D dense facial meshes corresponding to the user-constraints can be synthesized effectively.
  • Keywords
    belief networks; graph theory; image processing; 3D facial expression editing; basis expression probabilities; dynamic graph model; hierarchical dynamic Bayesian network; local facial features; sparse feature mesh; Bayesian methods; Context modeling; Facial features; Facial muscles; Hidden Markov models; Humans; Laboratories; Network synthesis; Shape; Speech synthesis; Bayesian network; Expression editing;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Multimedia and Expo, 2009. ICME 2009. IEEE International Conference on
  • Conference_Location
    New York, NY
  • ISSN
    1945-7871
  • Print_ISBN
    978-1-4244-4290-4
  • Electronic_ISBN
    1945-7871
  • Type

    conf

  • DOI
    10.1109/ICME.2009.5202754
  • Filename
    5202754