• DocumentCode
    3136103
  • Title

    Generalized adaptive view-based appearance model: Integrated framework for monocular head pose estimation

  • Author

    Morency, Louis-Philippe ; Whitehill, Jacob ; Movellan, Javier

  • Author_Institution
    Inst. for Creative Technol., USC, Marina del Rey, CA
  • fYear
    2008
  • fDate
    17-19 Sept. 2008
  • Firstpage
    1
  • Lastpage
    8
  • Abstract
    Accurately estimating the person´s head position and orientation is an important task for a wide range of applications such as driver awareness and human-robot interaction. Over the past two decades, many approaches have been suggested to solve this problem, each with its own advantages and disadvantages. In this paper, we present a probabilistic framework called generalized adaptive viewbased appearance model (GAVAM) which integrates the advantages from three of these approaches: (1) the automatic initialization and stability of static head pose estimation, (2) the relative precision and user-independence of differential registration, and (3) the robustness and bounded drift of keyframe tracking. In our experiments, we show how the GAVAM model can be used to estimate head position and orientation in real-time using a simple monocular camera. Our experiments on two previously published datasets show that the GAVAM framework can accurately track for a long period of time (>2 minutes) with an average accuracy of 3.5deg and 0.75 in with an inertial sensor and a 3D magnetic sensor.
  • Keywords
    pose estimation; probability; generalized adaptive view-based appearance model; monocular head pose estimation; probabilistic framework; Active appearance model; Cameras; Detectors; Face detection; Intelligent sensors; Magnetic heads; Magnetic sensors; Robust stability; Tracking; Uncertainty;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Automatic Face & Gesture Recognition, 2008. FG '08. 8th IEEE International Conference on
  • Conference_Location
    Amsterdam
  • Print_ISBN
    978-1-4244-2153-4
  • Electronic_ISBN
    978-1-4244-2154-1
  • Type

    conf

  • DOI
    10.1109/AFGR.2008.4813429
  • Filename
    4813429