Title :
Incremental communication for multilayer neural networks: error analysis
Author :
Ghorbani, Ali A. ; Bhavsar, Virendrakumar C.
Author_Institution :
Fac. of Comput. Eng., Iran Univ. of Sci. & Technol., Tehran, Iran
fDate :
1/1/1998 12:00:00 AM
Abstract :
Artificial neural networks (ANNs) involve a large amount of internode communications. To reduce the communication cost as well as the time of learning process in ANNs, we earlier proposed (1995) an incremental internode communication method. In the incremental communication method, instead of communicating the full magnitude of the output value of a node, only the increment or decrement to its previous value is sent to a communication link. In this paper, the effects of the limited precision incremental communication method on the convergence behavior and performance of multilayer neural networks are investigated. The nonlinear aspects of representing the incremental values with reduced (limited) precision for the commonly used error backpropagation training algorithm are analyzed. It is shown that the nonlinear effect of small perturbations in the input(s)/output of a node does not cause instability. The analysis is supported by simulation studies of two problems. The simulation results demonstrate that the limited precision errors are bounded and do not seriously affect the convergence of multilayer neural networks
Keywords :
backpropagation; convergence; error analysis; feedforward neural nets; multilayer perceptrons; convergence; error analysis; error backpropagation; finite precision computation; incremental communication; internode communications; learning process; multilayer neural networks; multilayer perceptrons; nonlinear effect; perturbations; Algorithm design and analysis; Artificial neural networks; Backpropagation algorithms; Computational modeling; Convergence; Costs; Degradation; Error analysis; Multi-layer neural network; Neural networks;
Journal_Title :
Neural Networks, IEEE Transactions on