• DocumentCode
    1031900
  • Title

    Improving generalization performance using double backpropagation

  • Author

    Drucker, Harris ; Le Cun, Yann

  • Author_Institution
    AT&T Bell Lab., West Long Branch, NJ, USA
  • Volume
    3
  • Issue
    6
  • fYear
    1992
  • fDate
    11/1/1992 12:00:00 AM
  • Firstpage
    991
  • Lastpage
    997
  • Abstract
    In order to generalize from a training set to a test set, it is desirable that small changes in the input space of a pattern do not change the output components. This can be done by forcing this behavior as part of the training algorithm. This is done in double backpropagation by forming an energy function that is the sum of the normal energy term found in backpropagation and an additional term that is a function of the Jacobian. Significant improvement is shown with different architectures and different test sets, especially with architectures that had previously been shown to have very good performance when trained using backpropagation. It is shown that double backpropagation, as compared to backpropagation, creates weights that are smaller, thereby causing the output of the neurons to spend more time in the linear region
  • Keywords
    backpropagation; neural nets; double backpropagation; energy function; generalization performance; learning; neural nets; training algorithm; Backpropagation algorithms; Jacobian matrices; Neural networks; Neurons; Signal to noise ratio; Testing;
  • fLanguage
    English
  • Journal_Title
    Neural Networks, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    1045-9227
  • Type

    jour

  • DOI
    10.1109/72.165600
  • Filename
    165600