• DocumentCode
    288324
  • Title

    Improving generalization performance by information minimization

  • Author

    Kamimura, Ryotaro ; Takagi, Toshiyuki ; Nakanishi, Shohachiro

  • Author_Institution
    Inf. Sci. Lab., Tokai Univ., Kanagawa, Japan
  • Volume
    1
  • fYear
    1994
  • fDate
    27 Jun-2 Jul 1994
  • Firstpage
    143
  • Abstract
    In this paper, we attempt to show that the information stored in networks must be as small as possible for the improvement of the generalization performance under the condition that the networks can produce targets with appropriate accuracy. The information is defined by the difference between maximum entropy or uncertainty and observed entropy. Borrowing a definition of fuzzy entropy, the uncertainty function is defined for the internal representation and represented by the equation: -υi log υi-(1-υi ) log (1-υi), where υi is a hidden unit activity. After having formulated an update rule for the minimization of the information, we applied the method to a problem of language acquisition: the inference of the past tense forms of regular verbs. Experimental results confirmed that by our method, the information was significantly decreased and the generalization performance was greatly improved
  • Keywords
    generalisation (artificial intelligence); grammars; inference mechanisms; information theory; knowledge acquisition; maximum entropy methods; minimisation; natural languages; neural nets; uncertainty handling; fuzzy entropy; generalization performance improvement; hidden unit activity; inference; information minimization; language acquisition; maximum entropy; neural nets; observed entropy; past tense; uncertainty function; verbs; Entropy; Equations; Information science; Laboratories; Minimization methods; Uncertainty;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks, 1994. IEEE World Congress on Computational Intelligence., 1994 IEEE International Conference on
  • Conference_Location
    Orlando, FL
  • Print_ISBN
    0-7803-1901-X
  • Type

    conf

  • DOI
    10.1109/ICNN.1994.374153
  • Filename
    374153