• Title of article

    Sparse coding for layered neural networks

  • Author/Authors

    Katsuki Katayama، نويسنده , , Yasuo Sakata، نويسنده , , Tsuyoshi Horiguchi، نويسنده ,

  • Issue Information
    روزنامه با شماره پیاپی سال 2002
  • Pages
    15
  • From page
    532
  • To page
    546
  • Abstract
    We investigate storage capacity of two types of fully connected layered neural networks with sparse coding when binary patterns are embedded into the networks by a Hebbian learning rule. One of them is a layered network, in which a transfer function of even layers is different from that of odd layers. The other is a layered network with intra-layer connections, in which the transfer function of inter-layer is different from that of intra-layer, and inter-layered neurons and intra-layered neurons are updated alternately. We derive recursion relations for order parameters by means of the signal-to-noise ratio method, and then apply the self-control threshold method proposed by Dominguez and Bollé to both layered networks with monotonic transfer functions. We find that a critical value αC of storage capacity is about 0.11a ln a−1 (a 1) for both layered networks, where a is a neuronal activity. It turns out that the basin of attraction is larger for both layered networks when the self-control threshold method is applied
  • Journal title
    Physica A Statistical Mechanics and its Applications
  • Serial Year
    2002
  • Journal title
    Physica A Statistical Mechanics and its Applications
  • Record number

    867846