DocumentCode
706077
Title
Multimodal fusion for cued Speech language recognition
Author
Argyropoulos, Savvas ; Tzovaras, Dimitrios ; Strintzis, Michael G.
Author_Institution
Electr. & Comput. Eng. Dept., Aristotle Univ. of Thessaloniki, Thessaloniki, Greece
fYear
2007
fDate
3-7 Sept. 2007
Firstpage
1289
Lastpage
1293
Abstract
In this paper, a novel method for Cued Speech language recognition is proposed. A multimodal processing framework is developed for the efficient fusion of the modalities under consideration. The robust feature extraction from the audio, lip shape and gesture modalities is also discussed. Moreover, the inherent correlation among the signals is modelled using a modified Coupled Hidden Markov Model. The contribution of each modality to the fused decision is modified by assessing its reliability and assigning a weighting factor to improve the accuracy of the recognized words. The method is experimentally evaluated and is shown to outperform methods that rely only on the perceivable information to infer the transmitted messages.
Keywords
feature extraction; hidden Markov models; speech recognition; audio modality; cued speech language recognition; gesture modality; lip shape modality; modified coupled hidden Markov model; multimodal fusion; multimodal processing framework; robust feature extraction; weighting factor; Feature extraction; Gesture recognition; Hidden Markov models; Reliability; Shape; Speech; Speech recognition;
fLanguage
English
Publisher
ieee
Conference_Titel
Signal Processing Conference, 2007 15th European
Conference_Location
Poznan
Print_ISBN
978-839-2134-04-6
Type
conf
Filename
7099013
Link To Document