DocumentCode :
2868676
Title :
Using an MDL-based cost function with neural networks
Author :
Lappalainen, Harri
Author_Institution :
Neural Networks Res. Centre, Helsinki Univ. of Technol., Hut, Finland
Volume :
3
fYear :
1998
fDate :
4-9 May 1998
Firstpage :
2384
Abstract :
The minimum description length (MDL) principle is an information theoretically based method to learn models from data. This paper presents an approach to efficiently use an MDL-based cost function with neural networks. As usual, the cost function can be used to adapt the parameters in the network, but it can also include terms to measure the complexity of the structure of the network and can thus be applied to determine the optimal structure. The basic idea is to convert a conventional neural network such that each parameter and each output of the neurons is assigned a means and a variance. This greatly simplifies the computation of the description length and its gradient with respect to the parameters, which can then be adapted using the standard gradient descent method
Keywords :
data compression; data structures; encoding; learning (artificial intelligence); multilayer perceptrons; cost function; data compression; data structures; encoding; gradient descent method; information theory; learning; minimum description length; multilayer perceptrons; neural networks; Bandwidth; Bayesian methods; Cost function; Electronic mail; Length measurement; Neural networks; Neurons; Pattern recognition; Testing;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks Proceedings, 1998. IEEE World Congress on Computational Intelligence. The 1998 IEEE International Joint Conference on
Conference_Location :
Anchorage, AK
ISSN :
1098-7576
Print_ISBN :
0-7803-4859-1
Type :
conf
DOI :
10.1109/IJCNN.1998.687235
Filename :
687235
Link To Document :
بازگشت