Title :
Asynchronous stream modeling for large vocabulary audio-visual speech recognition
Author :
Luettin, Juergen ; Potamianos, Gerasimos ; Neti, Chalapathy
Author_Institution :
Ascom Systec AG, Switzerland
Abstract :
Addresses the problem of audio-visual information fusion to provide highly robust speech recognition. We investigate methods that make different assumptions about asynchrony and conditional dependence across streams and propose a technique based on composite HMMs that can account for stream asynchrony and different levels of information integration. We show how these models can be trained jointly based on maximum likelihood estimation. Experiments, performed for a speaker-independent large vocabulary continuous speech recognition task and different integration methods, show that best performance is obtained by asynchronous stream integration. This system reduces the error rate at a 8.5 dB SNR with additive speech "babble" noise by 27 % relative over audio-only models and by 12 % relative over traditional audio-visual models using concatenative feature fusion
Keywords :
audio signal processing; feature extraction; hidden Markov models; sensor fusion; speech recognition; video signal processing; 8.5 dB; additive speech babble noise; asynchronous stream modeling; asynchrony; audio-visual information fusion; conditional dependence; highly robust speech recognition; large vocabulary audio-visual speech recognition; maximum likelihood estimation; speaker-independent speech recognition; Additive noise; Error analysis; Hidden Markov models; Maximum likelihood estimation; Robustness; Signal to noise ratio; Speech enhancement; Speech recognition; Streaming media; Vocabulary;
Conference_Titel :
Acoustics, Speech, and Signal Processing, 2001. Proceedings. (ICASSP '01). 2001 IEEE International Conference on
Conference_Location :
Salt Lake City, UT
Print_ISBN :
0-7803-7041-4
DOI :
10.1109/ICASSP.2001.940794