• Title of article

    Learning capability and storage capacity of two-hidden-layer feedforward networks

  • Author/Authors

    Huang، Guang-Bin نويسنده ,

  • Issue Information
    روزنامه با شماره پیاپی سال 2003
  • Pages
    -273
  • From page
    274
  • To page
    0
  • Abstract
    In this paper, we present a new class of quasi-Newton methods for an effective learning in large multilayer perceptron (MLP)-networks. The algorithms introduced in this work, named LQN, utilize an iterative scheme of a generalized BFGS-type method, involving a suitable family of matrix algebras L. The main advantages of these innovative methods are based upon the fact that they have an O(nlogn) complexity per step and that they require O(n) memory allocations. Numerical experiences, performed on a set of standard benchmarks of MLP-networks, show the competitivity of the LQN methods, especially for large values of n.
  • Keywords
    two-hidden-layer feedforward networks (TLFNs) , Storage capacity , Learning capability , neural-network modularity
  • Journal title
    IEEE TRANSACTIONS ON NEURAL NETWORKS
  • Serial Year
    2003
  • Journal title
    IEEE TRANSACTIONS ON NEURAL NETWORKS
  • Record number

    62809