DocumentCode :
572853
Title :
Strategies for constructive neural networks and its application to regression models
Author :
Jifu Nong
Author_Institution :
Coll. of Sci., Guangxi Univ. for Nat., Nanning, China
fYear :
2012
fDate :
24-26 Aug. 2012
Firstpage :
197
Lastpage :
201
Abstract :
Regression problem is an important application area for neural networks (NNs). Among a large number of existing NN architectures, the feedforward NN (FNN) paradigm is one of the most widely used structures. Although one-hidden-layer feedforward neural networks (OHL-FNNs) have simple structures, they possess interesting representational and learning capabilities. In this paper, we are interested particularly in incremental constructive training of OHL-FNNs. In the proposed incremental constructive training schemes for an OHL-FNN, input-side training and output-side training may be separated in order to reduce the training time. A new technique is proposed to scale the error signal during the constructive learning process to improve the input-side training efficiency and to obtain better generalization performance. Two pruning methods for removing the input-side redundant connections have also been applied. Numerical simulations demonstrate the potential and advantages of the proposed strategies when compared to other existing techniques in the literature.
Keywords :
feedforward neural nets; numerical analysis; regression analysis; FNN; OHL-FNNs; constructive neural networks; feedforward NN; numerical simulations; one hidden layer feedforward neural networks; regression models; regression problem; Artificial neural networks; Hafnium; Training; constructive neural networks; network pruning; regression models; training strategy;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Computer Science and Information Processing (CSIP), 2012 International Conference on
Conference_Location :
Xi´an, Shaanxi
Print_ISBN :
978-1-4673-1410-7
Type :
conf
DOI :
10.1109/CSIP.2012.6308828
Filename :
6308828
Link To Document :
بازگشت