DocumentCode :
2286743
Title :
Exploiting multimodal data fusion in robust speech recognition
Author :
Heracleous, Panikos ; Badin, Pierre ; Bailly, Gérard ; Hagita, Norihiro
Author_Institution :
ATR, Intell. Robot. & Commun. Labs., Japan
fYear :
2010
fDate :
19-23 July 2010
Firstpage :
568
Lastpage :
572
Abstract :
This article introduces automatic speech recognition based on Electro-Magnetic Articulography (EMA). Movements of the tongue, lips, and jaw are tracked by an EMA device, which are used as features to create Hidden Markov Models (HMM) and recognize speech only from articulation, that is, without any audio information. Also, automatic phoneme recognition experiments are conducted to examine the contribution of the EMA parameters to robust speech recognition. Using feature fusion, multistream HMM fusion, and late fusion methods, noisy audio speech has been integrated with EMA speech and recognition experiments have been conducted. The achieved results show that the integration of the EMA parameters significantly increases an audio speech recognizer´s accuracy, in noisy environments.
Keywords :
hidden Markov models; sensor fusion; speech recognition; articulation; audio information; automatic phoneme recognition; electro-magnetic articulography; feature fusion; hidden Markov model; late fusion methods; multimodal data fusion; multistream HMM fusion; noisy audio speech; robust speech recognition; Accuracy; Coils; Hidden Markov models; Noise measurement; Speech; Speech recognition; Tongue;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Multimedia and Expo (ICME), 2010 IEEE International Conference on
Conference_Location :
Suntec City
ISSN :
1945-7871
Print_ISBN :
978-1-4244-7491-2
Type :
conf
DOI :
10.1109/ICME.2010.5583086
Filename :
5583086
Link To Document :
بازگشت