DocumentCode :
313633
Title :
Modelling the perceptual separation of concurrent vowels with a network of neural oscillators
Author :
Brown, Guy J. ; Wang, DeLiang
Author_Institution :
Dept. of Comput. Sci., Sheffield Univ., UK
Volume :
1
fYear :
1997
fDate :
9-12 Jun 1997
Firstpage :
569
Abstract :
The ability of listeners to identify two simultaneously presented vowels is improved by introducing a difference in fundamental frequency between the vowels. We propose an explanation for this phenomenon in the form of a computational model of concurrent sound segregation, which is motivated by neurophysiological evidence of oscillatory firing activity in the higher auditory system. In the model, the perceptual grouping of auditory peripheral channels is coded by synchronised oscillations in a neural oscillator network. Computer simulations confirm that the model qualitatively matches the double vowel identification performance of human listeners
Keywords :
acoustic signal processing; auditory evoked potentials; biology computing; digital simulation; neural nets; neurophysiology; physiological models; speech recognition; auditory peripheral channels; concurrent sound segregation; concurrent vowels; double vowel identification performance; fundamental frequency; higher auditory system; human listeners; neural oscillators; oscillatory firing activity; perceptual grouping; perceptual separation; synchronised oscillations; Auditory system; Band pass filters; Channel bank filters; Ear; Frequency synchronization; Hair; Image analysis; Oscillators; Signal processing; Speech recognition;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks,1997., International Conference on
Conference_Location :
Houston, TX
Print_ISBN :
0-7803-4122-8
Type :
conf
DOI :
10.1109/ICNN.1997.611732
Filename :
611732
Link To Document :
بازگشت