• DocumentCode
    982805
  • Title

    Backpropagation uses prior information efficiently

  • Author

    Barnard, Etienne ; Botha, Elizabeth C.

  • Author_Institution
    Dept. of Comput. Sci. & Eng., Oregon Graduate Inst., Portland, OR, USA
  • Volume
    4
  • Issue
    5
  • fYear
    1993
  • fDate
    9/1/1993 12:00:00 AM
  • Firstpage
    794
  • Lastpage
    802
  • Abstract
    The ability of neural net classifiers to deal with a priori information is investigated. For this purpose, backpropagation classifiers are trained with data from known distributions with variable a priori probabilities, and their performance on separate test sets is evaluated. It is found that backpropagation employs a priori information in a slightly suboptimal fashion, but this does not have serious consequences on the performance of the classifier. Furthermore, it is found that the inferior generalization that results when an excessive number of network parameters are used can (partially) be ascribed to this suboptimality
  • Keywords
    backpropagation; generalisation (artificial intelligence); neural nets; pattern recognition; probability; a priori information; backpropagation; generalization; neural net classifiers; probabilities; Backpropagation; Computer science; Humans; Information resources; Least squares approximation; Neural networks; Pattern recognition; Speech; Springs; Testing;
  • fLanguage
    English
  • Journal_Title
    Neural Networks, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    1045-9227
  • Type

    jour

  • DOI
    10.1109/72.248457
  • Filename
    248457