DocumentCode :
375512
Title :
Recurrent neural network design for temporal sequence learning
Author :
Kwan, H.K. ; Yan, J.
Author_Institution :
Dept. of Electr. & Comput. Eng., Windsor Univ., Ont., Canada
Volume :
2
fYear :
2000
fDate :
2000
Firstpage :
832
Abstract :
Presents two designs of a recurrent neural network with 1st and 2nd order self-feedback at the hidden layer. The first design is based on a gradient descent algorithm and the second design is based on a genetic algorithm (GA). The simulation results of the single hidden layer network and those of a single hidden layer feedforward neural network for learning 50 commands of up to 3 words and 24 phone numbers of 10 digits are presented. Results indicate that the GA-based dynamic recurrent neural network is best in both convergence and error performance
Keywords :
feedforward neural nets; genetic algorithms; gradient methods; learning (artificial intelligence); recurrent neural nets; speech recognition; convergence; error performance; feedforward neural network; genetic algorithm; gradient descent algorithm; hidden layer; recurrent neural network; self-feedbacks; speech recognition; temporal sequence learning; Algorithm design and analysis; Computer networks; Convergence; Design engineering; Feedforward neural networks; Feedforward systems; Genetic algorithms; Neural networks; Neurons; Recurrent neural networks;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Circuits and Systems, 2000. Proceedings of the 43rd IEEE Midwest Symposium on
Conference_Location :
Lansing, MI
Print_ISBN :
0-7803-6475-9
Type :
conf
DOI :
10.1109/MWSCAS.2000.952884
Filename :
952884
Link To Document :
بازگشت