DocumentCode
288328
Title
Method of digging tunnels horizontally into the error hypersurface to speed up training and to escape from local minima
Author
Liang, Xun ; Xia, Shaowei ; Du, Jihong
Author_Institution
Dept. of Autom., Tsinghua Univ., Beijing, China
Volume
1
fYear
1994
fDate
27 Jun-2 Jul 1994
Firstpage
181
Abstract
In this paper, a general compressing method is reviewed systematically at first. Then the idea and steps of digging horizontally into the error hypersurface are presented, as well as an example. Since there exists serious and complex nonlinearity in the error hypersurface, training by gradient descending techniques is often too slow when it is on a plateau and has the risk of trapping into local minima. Digging tunnels into the error hypersurface by means of rotation transformation will lead to the plateau to speed up the training or skip from local minima
Keywords
data compression; learning (artificial intelligence); multilayer perceptrons; transforms; compressing method; error hypersurface; gradient descending method; hidden neuron; learning speed; local minima; multilayer perceptron; neural nets; rotation transformation; tunnel digging; Multilayer perceptrons; Neurons;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks, 1994. IEEE World Congress on Computational Intelligence., 1994 IEEE International Conference on
Conference_Location
Orlando, FL
Print_ISBN
0-7803-1901-X
Type
conf
DOI
10.1109/ICNN.1994.374159
Filename
374159
Link To Document