DocumentCode :
2704057
Title :
Double backpropagation increasing generalization performance
Author :
Drucker, Harris ; Cun, Yann Le
Author_Institution :
AT&T Bell Lab., Holmdel, NJ, USA
fYear :
1991
fDate :
8-14 Jul 1991
Firstpage :
145
Abstract :
One test of a training algorithm is how well the algorithm generalizes from the training data to the test data. It is shown that a training algorithm termed double back-propagation improves generalization by simultaneously minimizing the normal energy term found in back-propagation and an additional energy term that is related to the sum of the squares of the input derivatives (gradients). In normal back-propagation training, minimizing the energy function tends to push the input gradient to zero. However, this is not always possible. Double back-propagation explicitly pushes the input gradients to zero, making the minimum broader, and increases the generalization on the test data. The authors show the improvement over normal back-propagation on four candidate architectures with a training set of 320 handwritten numbers and a test set of size 180
Keywords :
learning systems; neural nets; double back-propagation; energy term minimization; neural nets; training algorithm; Backpropagation algorithms; Ear; Educational institutions; Equations; Neurons; Testing; Training data; Transfer functions;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1991., IJCNN-91-Seattle International Joint Conference on
Conference_Location :
Seattle, WA
Print_ISBN :
0-7803-0164-1
Type :
conf
DOI :
10.1109/IJCNN.1991.155328
Filename :
155328
Link To Document :
بازگشت