DocumentCode :
1268527
Title :
Convergence of Cyclic and Almost-Cyclic Learning With Momentum for Feedforward Neural Networks
Author :
Wang, Jian ; Yang, Jie ; Wu, Wei
Author_Institution :
Sch. of Math. Sci., Dalian Univ. of Technol., Dalian, China
Volume :
22
Issue :
8
fYear :
2011
Firstpage :
1297
Lastpage :
1306
Abstract :
Two backpropagation algorithms with momentum for feedforward neural networks with a single hidden layer are considered. It is assumed that the training samples are supplied to the network in a cyclic or an almost-cyclic fashion in the learning procedure, i.e., in each training cycle, each sample of the training set is supplied in a fixed or a stochastic order respectively to the network exactly once. A restart strategy for the momentum is adopted such that the momentum coefficient is set to zero at the beginning of each training cycle. Corresponding weak and strong convergence results are then proved, indicating that the gradient of the error function goes to zero and the weight sequence goes to a fixed point, respectively. The convergence conditions on the learning rate, the momentum coefficient, and the activation functions are much relaxed compared with those of the existing results.
Keywords :
backpropagation; convergence; feedforward neural nets; backpropagation algorithm; cyclic learning; error function gradient; feedforward neural networks; learning rate; momentum coefficient; restart strategy; stochastic order; Artificial neural networks; Convergence; Feedforward neural networks; Gradient methods; Neurons; Stability analysis; Training; Almost-cyclic; backpropagation; convergence; cyclic; feedforward neural networks; momentum; Artificial Intelligence; Neural Networks (Computer); Pattern Recognition, Automated;
fLanguage :
English
Journal_Title :
Neural Networks, IEEE Transactions on
Publisher :
ieee
ISSN :
1045-9227
Type :
jour
DOI :
10.1109/TNN.2011.2159992
Filename :
5948413
Link To Document :
بازگشت