DocumentCode :
1646679
Title :
Symmetry of backpropagation and chain rule
Author :
Pacut, Andrzej
Author_Institution :
Warsaw Univ. of Technol., Poland
Volume :
1
fYear :
2002
fDate :
6/24/1905 12:00:00 AM
Firstpage :
530
Lastpage :
534
Abstract :
Gradient backpropagation, as a method of computing derivatives of composite functions, is commonly understood as a version of the chain rule. We show that this is not true, and both methods are in a sense opposite. As for the chain rule one needs derivatives with respect to all variables that influence a given intermediate variable, the backpropagation calls for derivatives of all variables that are influenced by the present variable. Knowing this, the derivation of the gradient even for complicated neural networks is almost trivial. In a matrix form, both methods differ in the order of matrix multiplication. The use of the chain rule is almost automatic, while the use of the backpropagation can be automatic as an equivalent alternative version of the derivative calculation for composite functions
Keywords :
backpropagation; differential equations; matrix algebra; multilayer perceptrons; chain rule; composite functions; gradient backpropagation; matrix multiplication; multidimensional ordered systems; multilayer perceptron; neural networks; Algebra; Backpropagation algorithms; Equations; Gradient methods; Mirrors; Multilayer perceptrons; Neural networks; Reflection;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2002. IJCNN '02. Proceedings of the 2002 International Joint Conference on
Conference_Location :
Honolulu, HI
ISSN :
1098-7576
Print_ISBN :
0-7803-7278-6
Type :
conf
DOI :
10.1109/IJCNN.2002.1005528
Filename :
1005528
Link To Document :
بازگشت