• DocumentCode
    830098
  • Title

    Automatic Temporal Segment Detection and Affect Recognition From Face and Body Display

  • Author

    Gunes, Hatice ; Piccardi, Massimo

  • Author_Institution
    Univ. of Technol., Sydney, NSW
  • Volume
    39
  • Issue
    1
  • fYear
    2009
  • Firstpage
    64
  • Lastpage
    84
  • Abstract
    Psychologists have long explored mechanisms with which humans recognize other humans´ affective states from modalities, such as voice and face display. This exploration has led to the identification of the main mechanisms, including the important role played in the recognition process by the modalities´ dynamics. Constrained by the human physiology, the temporal evolution of a modality appears to be well approximated by a sequence of temporal segments called onset, apex, and offset. Stemming from these findings, computer scientists, over the past 15 years, have proposed various methodologies to automate the recognition process. We note, however, two main limitations to date. The first is that much of the past research has focused on affect recognition from single modalities. The second is that even the few multimodal systems have not paid sufficient attention to the modalities´ dynamics: The automatic determination of their temporal segments, their synchronization to the purpose of modality fusion, and their role in affect recognition are yet to be adequately explored. To address this issue, this paper focuses on affective face and body display, proposes a method to automatically detect their temporal segments or phases, explores whether the detection of the temporal phases can effectively support recognition of affective states, and recognizes affective states based on phase synchronization/alignment. The experimental results obtained show the following: 1) affective face and body displays are simultaneous but not strictly synchronous; 2) explicit detection of the temporal phases can improve the accuracy of affect recognition; 3) recognition from fused face and body modalities performs better than that from the face or the body modality alone; and 4) synchronized feature-level fusion achieves better performance than decision-level fusion.
  • Keywords
    behavioural sciences computing; emotion recognition; face recognition; object detection; affect recognition; automatic temporal segment detection; body display; face recognition; temporal phases; Affect recognition; affective face and body display; phase synchronization; selective fusion; temporal segment detection; Algorithms; Artificial Intelligence; Facial Expression; Gestures; Humans; Image Processing, Computer-Assisted; Movement; Normal Distribution; Pattern Recognition, Automated; Recognition (Psychology); Time Factors; Video Recording;
  • fLanguage
    English
  • Journal_Title
    Systems, Man, and Cybernetics, Part B: Cybernetics, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    1083-4419
  • Type

    jour

  • DOI
    10.1109/TSMCB.2008.927269
  • Filename
    4595624