• DocumentCode
    567732
  • Title

    A training algorithm and stability analysis for recurrent neural networks

  • Author

    Xu, Zhao ; Song, Qing ; Wang, Danwei ; Fan, Haijin

  • Author_Institution
    Sch. of Electr. & Electron. Eng., Nanyang Technol. Univ., Singapore, Singapore
  • fYear
    2012
  • fDate
    9-12 July 2012
  • Firstpage
    2285
  • Lastpage
    2292
  • Abstract
    Training of recurrent neural networks (RNNs) introduces considerable computational complexities due to the need for gradient evaluations. How to get fast convergence speed and low computational complexity remains a challenging and open topic. Besides, the transient response of learning process of RNNs is a critical issue, especially for on-line applications. Conventional RNNs training algorithms such as the backpropagation through time (BPTT) and real-time recurrent learning (RTRL) have not adequately satisfied these requirements because they often suffer from slow convergence speed. If a large learning rate is chosen to improve performance, the training process may become unstable in terms of weight divergence. In this paper, a novel training algorithm of RNN, named robust recurrent simultaneous perturbation stochastic approximation (RRSPSA), is developed with a specially designed recurrent hybrid adaptive parameter and adaptive learning rates. RRSPSA is a powerful novel twin-engine simultaneous perturbation stochastic approximation (SPSA) type of RNN training algorithm. It utilizes specific designed three adaptive parameters to maximize training speed for recurrent training signal while exhibiting certain weight convergence properties with only two objective function measurements as the original SPSA algorithm. The RRSPSA is proved with guaranteed weight convergence and system stability in the sense of Lyapunov function. Computer simulations were carried out to demonstrate applicability of the theoretical results.
  • Keywords
    Lyapunov methods; approximation theory; computational complexity; convergence of numerical methods; learning (artificial intelligence); recurrent neural nets; stability; stochastic processes; Lyapunov function; RNN; RRSPSA; adaptive learning rates; computational complexities; convergence speed; gradient evaluations; objective function measurements; online applications; performance improvement; recurrent hybrid adaptive parameter; recurrent neural networks; recurrent training signal; robust recurrent simultaneous perturbation stochastic approximation; stability analysis; training algorithm; training speed maximization; transient response; twin-engine simultaneous perturbation stochastic approx- imation; weight convergence properties; weight divergence; Algorithm design and analysis; Approximation algorithms; Approximation methods; Convergence; Stability analysis; Training; Vectors; recurrent neural networks (RNNs); simultaneous perturbation stochastic approximation (SPSA) training; weight convergence and stability proofs;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Information Fusion (FUSION), 2012 15th International Conference on
  • Conference_Location
    Singapore
  • Print_ISBN
    978-1-4673-0417-7
  • Electronic_ISBN
    978-0-9824438-4-2
  • Type

    conf

  • Filename
    6290583