DocumentCode
3424869
Title
Convergence rates for a class of neural networks with logarithmic function
Author
Cao, Feilong ; Yuan, Yubo
Author_Institution
Inst. of Metrol. & Comput. Sci., China Jiliang Univ., Hangzhou, China
fYear
2009
fDate
17-19 Aug. 2009
Firstpage
27
Lastpage
32
Abstract
The aim of this paper is to estimate the approximation error which results from the method of feedforward neural networks (FNNs) with logarithmic sigmoidal function s(x) = (1 + e-x)-1. By means of an extending function approach, a class of FNNs with single hidden layer and the active function s(x) is constructed to approximate the continuous function defined on a compact interval. By using the modulus of continuity of function as metric, the rate of convergence of the FNNs is estimated. Also, a numerical examples for illustrating the theoretical results is given.
Keywords
convergence; recurrent neural nets; active function; approximation error estimation; convergence rate; extending function approach; feedforward neural network; logarithmic sigmoidal function; single hidden layer; Approximation error; Biological system modeling; Computational biology; Computer networks; Convergence; Demography; Feedforward neural networks; Logistics; Metrology; Neural networks;
fLanguage
English
Publisher
ieee
Conference_Titel
Granular Computing, 2009, GRC '09. IEEE International Conference on
Conference_Location
Nanchang
Print_ISBN
978-1-4244-4830-2
Type
conf
DOI
10.1109/GRC.2009.5255166
Filename
5255166
Link To Document