DocumentCode :
2722643
Title :
Training speed-up methods for neural networks applied to word recognition
Author :
Thierer, Gebhard ; Krause, Andreas ; Hackbarth, Heidi
Author_Institution :
SEL Alcatel, Stuttgart, Germany
fYear :
1991
fDate :
8-14 Jul 1991
Firstpage :
865
Abstract :
Two methods for speeding up the training of neural networks for word recognition are presented. The idea of the first method is to reduce the number of training patterns. The number of vocabulary repetitions can be cut down from seven to one if a pretrained network is used as the basis for further learning instead of a randomly initialized network. The second method does not need a pretrained network. Instead, training is performed alternatively with the entire training set and a subset thereof. This saves unnecessary backpropagation cycles for patterns that have already been learned. Depending on the data material, network training time is reduced at least by a factor of seven in the first case, and by a factor of two in the second case. Moreover, the error rate for a 100-word vocabulary can be lowered by one fourth by applying the second method
Keywords :
artificial intelligence; learning systems; neural nets; speech recognition; 100-word vocabulary; backpropagation cycles; error rate; learning; neural networks; training speed-up; vocabulary repetitions; word recognition; Artificial neural networks; Cepstrum; Databases; Error analysis; Multilayer perceptrons; Neural networks; Optimizing compilers; Radio access networks; Speech recognition; Vocabulary;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1991., IJCNN-91-Seattle International Joint Conference on
Conference_Location :
Seattle, WA
Print_ISBN :
0-7803-0164-1
Type :
conf
DOI :
10.1109/IJCNN.1991.155448
Filename :
155448
Link To Document :
بازگشت