• DocumentCode
    706077
  • Title

    Multimodal fusion for cued Speech language recognition

  • Author

    Argyropoulos, Savvas ; Tzovaras, Dimitrios ; Strintzis, Michael G.

  • Author_Institution
    Electr. & Comput. Eng. Dept., Aristotle Univ. of Thessaloniki, Thessaloniki, Greece
  • fYear
    2007
  • fDate
    3-7 Sept. 2007
  • Firstpage
    1289
  • Lastpage
    1293
  • Abstract
    In this paper, a novel method for Cued Speech language recognition is proposed. A multimodal processing framework is developed for the efficient fusion of the modalities under consideration. The robust feature extraction from the audio, lip shape and gesture modalities is also discussed. Moreover, the inherent correlation among the signals is modelled using a modified Coupled Hidden Markov Model. The contribution of each modality to the fused decision is modified by assessing its reliability and assigning a weighting factor to improve the accuracy of the recognized words. The method is experimentally evaluated and is shown to outperform methods that rely only on the perceivable information to infer the transmitted messages.
  • Keywords
    feature extraction; hidden Markov models; speech recognition; audio modality; cued speech language recognition; gesture modality; lip shape modality; modified coupled hidden Markov model; multimodal fusion; multimodal processing framework; robust feature extraction; weighting factor; Feature extraction; Gesture recognition; Hidden Markov models; Reliability; Shape; Speech; Speech recognition;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Signal Processing Conference, 2007 15th European
  • Conference_Location
    Poznan
  • Print_ISBN
    978-839-2134-04-6
  • Type

    conf

  • Filename
    7099013