DocumentCode :
2797462
Title :
Optimal Size of a Feedforward Neural Network: How Much does it Matter?
Author :
Wang, Lipo ; Quek, Hou Chai ; Tee, Keng Hoe ; Zhou, Nina ; Wan, Chunru
Author_Institution :
Coll. of Inf. Eng., Xiangtan Univ.
fYear :
2005
fDate :
23-28 Oct. 2005
Firstpage :
69
Lastpage :
69
Abstract :
In this paper, we attempt to answer the following question with systematic computer simulations: for the same validation error rate, does the size of a feedforward neural network matter? This is related to the so-called Occam´s Razor, that is, with all things being equal, the simplest solution is likely to work the best. Our simulation results indicate that for the same validation error rate, smaller networks do not tend to work better than larger networks, that is, Occam´s Razor does not seem to apply to feedforward neural networks. In fact, our results show no trend between network size and performance for a given validation error
Keywords :
feedforward neural nets; Occam´s Razor; feedforward neural network; validation error rate; Artificial neural networks; Computational modeling; Computer simulation; Educational institutions; Error analysis; Feedforward neural networks; Multilayer perceptrons; Neural networks; Neurons; Training data; Hidden neurons.; Learning; Neural networks; Occam’s Razor;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Autonomic and Autonomous Systems and International Conference on Networking and Services, 2005. ICAS-ICNS 2005. Joint International Conference on
Conference_Location :
Papeete, Tahiti
Print_ISBN :
0-7695-2450-8
Type :
conf
DOI :
10.1109/ICAS-ICNS.2005.72
Filename :
1559921
Link To Document :
بازگشت