DocumentCode
2787638
Title
Decision level combination of multiple modalities for recognition and analysis of emotional expression
Author
Metallinou, Angeliki ; Lee, Sungbok ; Narayanan, Shrikanth
Author_Institution
Dept. of Electr. Eng., Univ. of Southern California, Los Angeles, CA, USA
fYear
2010
fDate
14-19 March 2010
Firstpage
2462
Lastpage
2465
Abstract
Emotion is expressed and perceived through multiple modalities. In this work, we model face, voice and head movement cues for emotion recognition and we fuse classifiers using a Bayesian framework. The facial classifier is the best performing followed by the voice and head classifiers and the multiple modalities seem to carry complementary information, especially for happiness. Decision fusion significantly increases the average total unweighted accuracy, from 55% to about 62%. Overall, we achieve average accuracy on the order of 65-75% for emotional states and 30-40% for neutral state using a large multi-speaker, multimodal database. Performance analysis for the case of anger and neutrality suggests a positive correlation between the number of classifiers that performed well and the perceptual salience of the expressed emotion.
Keywords
Bayes methods; emotion recognition; face recognition; Bayesian framework; decision level combination; emotional expression analysis; emotional expression recognition; multimodal database; multiple modalities; performance analysis; Bayesian methods; Databases; Emotion recognition; Face detection; Fuses; Head; Hidden Markov models; Performance analysis; Robustness; Speech; Bayesian Information Fusion; Hidden Markov Model; Multimodal Emotion Recognition; Perceptual Salience;
fLanguage
English
Publisher
ieee
Conference_Titel
Acoustics Speech and Signal Processing (ICASSP), 2010 IEEE International Conference on
Conference_Location
Dallas, TX
ISSN
1520-6149
Print_ISBN
978-1-4244-4295-9
Electronic_ISBN
1520-6149
Type
conf
DOI
10.1109/ICASSP.2010.5494890
Filename
5494890
Link To Document