• DocumentCode
    2260299
  • Title

    The inefficiency of batch training for large training sets

  • Author

    Wilson, D. Randall ; Martinez, Tony R.

  • Author_Institution
    Fonix Corp., USA
  • Volume
    2
  • fYear
    2000
  • fDate
    2000
  • Firstpage
    113
  • Abstract
    Multilayer perceptrons are often trained using error backpropagation (BP). BP training can be done in either a batch or continuous manner. Claims have frequently been made that batch training is faster and/or more “correct” than continuous training because it uses a better approximation of the true gradient for its weight updates. These claims are often supported by empirical evidence on very small data sets. These claims are untrue, however, for large training sets. This paper explains why batch training is much slower than continuous training for large training sets. Various levels of semi-batch training used on a 20,000-instance speech recognition task show a roughly linear increase in training time required with an increase in batch size
  • Keywords
    backpropagation; multilayer perceptrons; speech recognition; batch learning; error backpropagation; multilayer perceptrons; speech recognition; weight updates; Backpropagation; Computer networks; Equations; Frequency; Multilayer perceptrons; Speech recognition;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks, 2000. IJCNN 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on
  • Conference_Location
    Como
  • ISSN
    1098-7576
  • Print_ISBN
    0-7695-0619-4
  • Type

    conf

  • DOI
    10.1109/IJCNN.2000.857883
  • Filename
    857883