DocumentCode
572853
Title
Strategies for constructive neural networks and its application to regression models
Author
Jifu Nong
Author_Institution
Coll. of Sci., Guangxi Univ. for Nat., Nanning, China
fYear
2012
fDate
24-26 Aug. 2012
Firstpage
197
Lastpage
201
Abstract
Regression problem is an important application area for neural networks (NNs). Among a large number of existing NN architectures, the feedforward NN (FNN) paradigm is one of the most widely used structures. Although one-hidden-layer feedforward neural networks (OHL-FNNs) have simple structures, they possess interesting representational and learning capabilities. In this paper, we are interested particularly in incremental constructive training of OHL-FNNs. In the proposed incremental constructive training schemes for an OHL-FNN, input-side training and output-side training may be separated in order to reduce the training time. A new technique is proposed to scale the error signal during the constructive learning process to improve the input-side training efficiency and to obtain better generalization performance. Two pruning methods for removing the input-side redundant connections have also been applied. Numerical simulations demonstrate the potential and advantages of the proposed strategies when compared to other existing techniques in the literature.
Keywords
feedforward neural nets; numerical analysis; regression analysis; FNN; OHL-FNNs; constructive neural networks; feedforward NN; numerical simulations; one hidden layer feedforward neural networks; regression models; regression problem; Artificial neural networks; Hafnium; Training; constructive neural networks; network pruning; regression models; training strategy;
fLanguage
English
Publisher
ieee
Conference_Titel
Computer Science and Information Processing (CSIP), 2012 International Conference on
Conference_Location
Xi´an, Shaanxi
Print_ISBN
978-1-4673-1410-7
Type
conf
DOI
10.1109/CSIP.2012.6308828
Filename
6308828
Link To Document