DocumentCode :
1543502
Title :
Neural network models of sensory integration for improved vowel recognition
Author :
Yuhas, Ben P. ; Goldstein, Moise H., Jr. ; Sejnowski, Terrence J. ; Jenkins, Robert E.
Author_Institution :
Bell Commum. Res., Morristown, NJ, USA
Volume :
78
Issue :
10
fYear :
1990
fDate :
10/1/1990 12:00:00 AM
Firstpage :
1658
Lastpage :
1668
Abstract :
It is demonstrated that multiple sources of speech information can be integrated at a subsymbolic level to improve vowel recognition. Feedforward and recurrent neural networks are trained to estimate the acoustic characteristics of a vocal tract from images of the speaker´s mouth. These estimates are then combined with the noise-degraded acoustic information, effectively increasing the signal-to-noise ratio and improving the recognition of these noise-degraded signals. Alternative symbolic strategies such as direct categorization of the visual signals into vowels are also presented. The performances of these neural networks compare favorably with human performance and with other pattern-matching and estimation techniques
Keywords :
neural nets; signal processing; speech recognition; neural networks; noise-degraded acoustic information; pattern-matching; sensory integration; speech recognition; vowel recognition; Acoustic noise; Automatic speech recognition; Degradation; Humans; Mouth; Neural networks; Signal to noise ratio; Speech enhancement; Speech recognition; Working environment noise;
fLanguage :
English
Journal_Title :
Proceedings of the IEEE
Publisher :
ieee
ISSN :
0018-9219
Type :
jour
DOI :
10.1109/5.58349
Filename :
58349
Link To Document :
بازگشت