DocumentCode :
706077
Title :
Multimodal fusion for cued Speech language recognition
Author :
Argyropoulos, Savvas ; Tzovaras, Dimitrios ; Strintzis, Michael G.
Author_Institution :
Electr. & Comput. Eng. Dept., Aristotle Univ. of Thessaloniki, Thessaloniki, Greece
fYear :
2007
fDate :
3-7 Sept. 2007
Firstpage :
1289
Lastpage :
1293
Abstract :
In this paper, a novel method for Cued Speech language recognition is proposed. A multimodal processing framework is developed for the efficient fusion of the modalities under consideration. The robust feature extraction from the audio, lip shape and gesture modalities is also discussed. Moreover, the inherent correlation among the signals is modelled using a modified Coupled Hidden Markov Model. The contribution of each modality to the fused decision is modified by assessing its reliability and assigning a weighting factor to improve the accuracy of the recognized words. The method is experimentally evaluated and is shown to outperform methods that rely only on the perceivable information to infer the transmitted messages.
Keywords :
feature extraction; hidden Markov models; speech recognition; audio modality; cued speech language recognition; gesture modality; lip shape modality; modified coupled hidden Markov model; multimodal fusion; multimodal processing framework; robust feature extraction; weighting factor; Feature extraction; Gesture recognition; Hidden Markov models; Reliability; Shape; Speech; Speech recognition;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Signal Processing Conference, 2007 15th European
Conference_Location :
Poznan
Print_ISBN :
978-839-2134-04-6
Type :
conf
Filename :
7099013
Link To Document :
بازگشت