Title :
Improving generalization performance by information minimization
Author :
Kamimura, Ryotaro ; Takagi, Toshiyuki ; Nakanishi, Shohachiro
Author_Institution :
Inf. Sci. Lab., Tokai Univ., Kanagawa, Japan
fDate :
27 Jun-2 Jul 1994
Abstract :
In this paper, we attempt to show that the information stored in networks must be as small as possible for the improvement of the generalization performance under the condition that the networks can produce targets with appropriate accuracy. The information is defined by the difference between maximum entropy or uncertainty and observed entropy. Borrowing a definition of fuzzy entropy, the uncertainty function is defined for the internal representation and represented by the equation: -υi log υi-(1-υi ) log (1-υi), where υi is a hidden unit activity. After having formulated an update rule for the minimization of the information, we applied the method to a problem of language acquisition: the inference of the past tense forms of regular verbs. Experimental results confirmed that by our method, the information was significantly decreased and the generalization performance was greatly improved
Keywords :
generalisation (artificial intelligence); grammars; inference mechanisms; information theory; knowledge acquisition; maximum entropy methods; minimisation; natural languages; neural nets; uncertainty handling; fuzzy entropy; generalization performance improvement; hidden unit activity; inference; information minimization; language acquisition; maximum entropy; neural nets; observed entropy; past tense; uncertainty function; verbs; Entropy; Equations; Information science; Laboratories; Minimization methods; Uncertainty;
Conference_Titel :
Neural Networks, 1994. IEEE World Congress on Computational Intelligence., 1994 IEEE International Conference on
Conference_Location :
Orlando, FL
Print_ISBN :
0-7803-1901-X
DOI :
10.1109/ICNN.1994.374153