Title :
When Does Online BP Training Converge?
Author :
Zong-Ben Xu ; Rui Zhang ; Wen-Feng Jing
Author_Institution :
Inst. for Inf. & Syst. Sci., Xi´an Jiaotong Univ., Xi´an, China
Abstract :
The backpropogation (BP) neural networks have been widely applied in scientific research and engineering. The success of the application, however, relies upon the convergence of the training procedure involved in the neural network learning. We settle down the convergence analysis issue through proving two fundamental theorems on the convergence of the online BP training procedure. One theorem claims that under mild conditions, the gradient sequence of the error function will converge to zero (the weak convergence), and another theorem concludes the convergence of the weight sequence defined by the procedure to a fixed value at which the error function attains its minimum (the strong convergence). The weak convergence theorem sharpens and generalizes the existing convergence analysis conducted before, while the strong convergence theorem provides new analysis results on convergence of the online BP training procedure. The results obtained reveal that with any analytic sigmoid activation function, the online BP training procedure is always convergent, which then underlies successful application of the BP neural networks.
Keywords :
backpropagation; convergence; gradient methods; neural nets; theorem proving; analytic sigmoid activation function; backpropogation neural networks; convergence analysis; convergence theorem; fundamental theorems proving; gradient sequence; neural network learning; online BP training; Backpropagation; Convergence; Feedforward neural networks; Gradient methods; H infinity control; Mathematics; Neural networks; Backpropagation (BP) neural networks; convergence analysis; online BP training procedure; Algorithms; Computer Simulation; Models, Theoretical; Neural Networks (Computer);
Journal_Title :
Neural Networks, IEEE Transactions on
DOI :
10.1109/TNN.2009.2025946