DocumentCode :
1748800
Title :
Relation between weight initialization of neural networks and pruning algorithms: case study on Mackey-Glass time series
Author :
Wan, Weishui ; Hirasawa, Kotaro ; Hu, Jinglu ; Murata, Junichi
Author_Institution :
Intelligent Control Lab., Kyushu Univ., Fukuoka, Japan
Volume :
3
fYear :
2001
fDate :
2001
Firstpage :
1750
Abstract :
The implementation of weight initialization is directly related to the convergence of learning algorithms. We made a case study on the Mackey-Glass time series problem in order to try to find some relations between weight initialization of neural networks and pruning algorithms. The pruning algorithm used in simulations is the Laplace regularizer method, that is, the backpropagation algorithm with Laplace regularizer added to the criterion function. Simulation results show that different kinds of initialization weight matrices display almost the same generalization ability when using the pruning algorithm, at least for the Mackey-Glass time series
Keywords :
Gaussian distribution; backpropagation; convergence; neural nets; time series; Laplace regularizer method; Mackey-Glass time series; backpropagation algorithm; convergence; criterion function; generalization ability; initialization weight matrices; learning algorithms; neural networks; pruning algorithms; weight initialization; Backpropagation algorithms; Computer aided software engineering; Concrete; Convergence; Data mining; Glass; Information science; Intelligent control; Laboratories; Neural networks;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2001. Proceedings. IJCNN '01. International Joint Conference on
Conference_Location :
Washington, DC
ISSN :
1098-7576
Print_ISBN :
0-7803-7044-9
Type :
conf
DOI :
10.1109/IJCNN.2001.938426
Filename :
938426
Link To Document :
بازگشت