Title :
Latent Mixture of Discriminative Experts
Author :
Ozkan, D. ; Morency, Louis-Philippe
Author_Institution :
Inst. for Creative Technol., Univ. of Southern California, Playa Vista, CA, USA
Abstract :
In this paper, we introduce a new model called Latent Mixture of Discriminative Experts which can automatically learn the temporal relationship between different modalities. Since, we train separate experts for each modality, LMDE is capable of improving the prediction performance even with limited amount of data. For model interpretation, we present a sparse feature ranking algorithm that exploits L1 regularization. An empirical evaluation is provided on the task of listener backchannel prediction (i.e., head nod). We introduce a new error evaluation metric called User-adaptive Prediction Accuracy that takes into account the difference in people´s backchannel responses. Our results confirm the importance of combining five types of multimodal features: lexical, syntactic structure, part-of-speech, visual and prosody. Latent Mixture of Discriminative Experts model outperforms previous approaches.
Keywords :
learning (artificial intelligence); sensor fusion; L1 regularization; LMDE model; empirical evaluation; error evaluation metric; latent mixture of discriminative experts model; lexical feature; listener backchannel prediction; modality learning; model interpretation; multimodal fusion; part-of-speech feature; prediction performance; prosody feature; sparse feature ranking algorithm; syntactic structure feature; user-adaptive prediction accuracy metric; visual feature; Accuracy; Computational modeling; Data models; Hidden Markov models; Measurement; Predictive models; Training; Backchannel feedback; evaluation metric; mixture of experts; multimodal integration; multimodal prediction models; sparse regularization;
Journal_Title :
Multimedia, IEEE Transactions on
DOI :
10.1109/TMM.2012.2229263