• DocumentCode
    3672090
  • Title

    Latent trees for estimating intensity of Facial Action Units

  • Author

    Sebastian Kaltwang;Sinisa Todorovic;Maja Pantic

  • Author_Institution
    Imperial College London, UK
  • fYear
    2015
  • fDate
    6/1/2015 12:00:00 AM
  • Firstpage
    296
  • Lastpage
    304
  • Abstract
    This paper is about estimating intensity levels of Facial Action Units (FAUs) in videos as an important step toward interpreting facial expressions. As input features, we use locations of facial landmark points detected in video frames. To address uncertainty of input, we formulate a generative latent tree (LT) model, its inference, and novel algorithms for efficient learning of both LT parameters and structure. Our structure learning iteratively builds LT by adding either a new edge or a new hidden node to LT, starting from initially independent nodes of observable features. A graph-edit operation that increases maximally the likelihood and minimally the model complexity is selected as optimal in each iteration. For FAU intensity estimation, we derive closed-form expressions of posterior marginals of all variables in LT, and specify an efficient bottom-up/top-down inference. Our evaluation on the benchmark DISFA and ShoulderPain datasets, in subject-independent setting, demonstrate that we outperform the state of the art, even under significant noise in facial landmarks. Effectiveness of our structure learning is demonstrated by probabilistically sampling meaningful facial expressions from the LT.
  • Keywords
    "Vegetation","Videos","Joints","Inference algorithms","Estimation","Training","Mathematical model"
  • Publisher
    ieee
  • Conference_Titel
    Computer Vision and Pattern Recognition (CVPR), 2015 IEEE Conference on
  • Electronic_ISBN
    1063-6919
  • Type

    conf

  • DOI
    10.1109/CVPR.2015.7298626
  • Filename
    7298626