DocumentCode :
149066
Title :
Emotion classification of speech using modulation features
Author :
Chaspari, Theodora ; Dimitriadis, Dimitrios ; Maragos, Petros
Author_Institution :
EE Dept., USC, Los Angeles, CA, USA
fYear :
2014
fDate :
1-5 Sept. 2014
Firstpage :
1552
Lastpage :
1556
Abstract :
Automatic classification of a speaker´s affective state is one of the major challenges in signal processing community, since it can improve Human-Computer interaction and give insights into the nature of emotions from psychology perspective. The amplitude and frequency control of sound production influences strongly the affective voice content. In this paper, we take advantage of the inherent speech modulations and propose the use of instant amplitude- and frequency-derived features for efficient emotion recognition. Our results indicate that these features can further increase the performance of the widely-used spectral-prosodic information, achieving improvements on two emotional databases, the Berlin Database of Emotional Speech and the recently collected Athens Emotional States Inventory.
Keywords :
emotion recognition; human computer interaction; speaker recognition; speech processing; Athens emotional states inventory; Berlin database of emotional speech; amplitude control; emotion recognition; frequency control; human-computer interaction; modulation features; signal processing; sound production; speaker classification; speech emotion classification; Databases; Emotion recognition; Feature extraction; Frequency modulation; Speech; Speech recognition; AM-FM features; Emotion classification; human-computer interaction; speech analysis;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Signal Processing Conference (EUSIPCO), 2014 Proceedings of the 22nd European
Conference_Location :
Lisbon
Type :
conf
Filename :
6952550
Link To Document :
بازگشت