• DocumentCode
    3497908
  • Title

    A new sensitivity-based pruning technique for feed-forward neural networks that improves generalization

  • Author

    Mrázová, Iveta ; Reitermanová, Zuzana

  • Author_Institution
    Dept. of Theor. Comput. Sci. & Math. Logic, Charles Univ. of Prague, Prague, Czech Republic
  • fYear
    2011
  • fDate
    July 31 2011-Aug. 5 2011
  • Firstpage
    2143
  • Lastpage
    2150
  • Abstract
    Multi-layer neural networks of the back-propagation type (MLP-networks) became a well-established tool used in various application areas. Reliable solutions require, however, also sufficient generalization capabilities of the formed networks and an easy interpretation of their function. These characteristics are strongly related to less sensitive networks with an optimized network structure. In this paper, we will introduce a new pruning technique called SCGSIR that is inspired by the fast method of scaled conjugate gradients (SCG) and sensitivity analysis. Network sensitivity inhibited during training impacts efficient optimization of network structure. Experiments performed so far yield promising results outperforming the reference techniques when considering both their ability to find networks with optimum architecture and improved generalization.
  • Keywords
    backpropagation; conjugate gradient methods; feedforward neural nets; MLP-network; SCGSIR; back-propagation type; feed-forward neural network; multilayer neural network; scaled conjugate gradient; sensitivity analysis; sensitivity-based pruning technique; Indexes; Neurons; Sensitivity analysis; Shape; Training; Transfer functions;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks (IJCNN), The 2011 International Joint Conference on
  • Conference_Location
    San Jose, CA
  • ISSN
    2161-4393
  • Print_ISBN
    978-1-4244-9635-8
  • Type

    conf

  • DOI
    10.1109/IJCNN.2011.6033493
  • Filename
    6033493