Title :
Cascade correlation: derivation of a more numerically stable update rule
Author_Institution :
Dept. of Comput. Sci., Stanford Univ., CA, USA
Abstract :
Discusses the weight update rule in the cascade correlation neural net learning algorithm. The weight update rule implements gradient descent optimization of the correlation between a new hidden unit´s output and the previous network´s error. The author presents a derivation of the gradient of the correlation function and shows that his resulting weight update rule results in slightly faster training. The author also shows that the new rule is mathematically equivalent to the one presented in the original cascade correlation paper and discusses numerical issues underlying the difference in performance. Since a derivation of the cascade correlation weight update rule was not published, this paper should be useful to those who wish to understand the rule
Keywords :
learning (artificial intelligence); neural nets; cascade correlation neural net; gradient descent optimization; learning algorithm; numerically stable update rule; weight update rule; Computer science; Equations; Error correction; Least squares methods; Network topology; Neural networks; Newton method; Optimization methods; Robots; Training data;
Conference_Titel :
Neural Networks, 1995. Proceedings., IEEE International Conference on
Conference_Location :
Perth, WA
Print_ISBN :
0-7803-2768-3
DOI :
10.1109/ICNN.1995.487581