• DocumentCode
    3542829
  • Title

    Comparison of training algorithms and network architectures

  • Author

    Wilamowski, Bogdan M.

  • Author_Institution
    Alabama Nano/Micro Sci. & Technol. Center, Auburn Univ., Auburn, AL, USA
  • fYear
    2013
  • fDate
    19-21 June 2013
  • Firstpage
    11
  • Lastpage
    11
  • Abstract
    Summary form only given. In the presentation major difficulties of designing neural networks are shown. It turn out that popular MLP (Multi Layer Perceptron) networks in most cases produces far from satisfactory results. Also, popular EBP (Error Back Propagation) algorithm is very slow and often is not capable to train best neural network architectures. Very powerful and fast LM (Levenberg- Marquardt) algorithm was unfortunately implemented only for MLP networks. Also, because a necessity of the inversion of the matrix, which size is proportional to number of patterns, the LM algorithm can be used only for small problems. However, the major frustration with neural networks occurs when too large neural networks are used and it is being trained with too small number of training patterns. Indeed, such networks, with excessive number of neurons, can be trained to very small errors, but these networks will respond very poorly for new patterns, which were not used for training. The most of frustrations with neural network can be eliminated when smaller, more effective, architectures are used and trained by newly developed NBN (Neuron-by-Neuron) algorithm. The methods of computational intelligence can be most successful but they have to be used with great care. It turns out that often popular training algorithms are not capable of tuning neural network to proper accuracy without losing generalization abilities. As a consequence, such system of computational intelligence may not work properly for cases which were not used in training. The comparison of different neural network architectures is given and it is shown how to develop and train close to optimal topologies.
  • Keywords
    backpropagation; generalisation (artificial intelligence); matrix algebra; multilayer perceptrons; EBP algorithm; LM algorithm; Levenberg-Marquardt algorithm; MLP network; NBN algorithm; Neuron-by-Neuron algorithm; computational intelligence; error backpropagation algorithm; generalization ability; matrix inversion; multilayer perceptron; neural network architectures; neural network design; training algorithms; Artificial intelligence; Biological neural networks; Computational intelligence; Computer architecture; Conferences; Neurons; Training;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Intelligent Engineering Systems (INES), 2013 IEEE 17th International Conference on
  • Conference_Location
    San Jose
  • Print_ISBN
    978-1-4799-0828-8
  • Type

    conf

  • DOI
    10.1109/INES.2013.6632792
  • Filename
    6632792