DocumentCode :
1935556
Title :
Phonetic to acoustic mapping using recurrent neural networks
Author :
Kumar, Vinod V. ; Ahalt, Stanley C. ; Krishnamurthy, Ashok K.
Author_Institution :
Dept. of Electr. Eng., Ohio State Univ., Columbus, OH, USA
fYear :
1991
fDate :
14-17 Apr 1991
Firstpage :
753
Abstract :
The application of artificial neural networks for phonetic-to-acoustic mapping is described. The specific task considered is that of mapping consonant-vowel-consonant (CVC) syllables to the corresponding formant values at different speech tempos. The performances of two different networks, the Elman recurrent network and a single hidden layer feedforward network, are compared. The results indicate that the recurrent network is able to generalize from the training set and produce valid formant contours for new CVC syllables that are not a part of the training set. It is shown that by choosing the proper input representation, the feedforward network is also capable of learning this mapping
Keywords :
neural nets; speech analysis and processing; speech recognition; speech synthesis; Elman recurrent network; consonant-vowel-consonant syllables; different speech tempos; formant contours; formant values; input representation; phonetic-to-acoustic mapping; recurrent neural networks; single hidden layer feedforward network; speech recognition; speech synthesis; training set; Acoustic applications; Artificial neural networks; Computer architecture; Delay effects; Network synthesis; Neural networks; Recurrent neural networks; Speech processing; Speech recognition; Speech synthesis;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech, and Signal Processing, 1991. ICASSP-91., 1991 International Conference on
Conference_Location :
Toronto, Ont.
ISSN :
1520-6149
Print_ISBN :
0-7803-0003-3
Type :
conf
DOI :
10.1109/ICASSP.1991.150450
Filename :
150450
Link To Document :
بازگشت