Title :
Nonlinear System Modeling With Random Matrices: Echo State Networks Revisited
Author :
Bai Zhang ; Miller, D.J. ; Yue Wang
Author_Institution :
Bradley Dept. of Electr. & Comput. Eng., Virginia Polytech. Inst. & State Univ., Arlington, VA, USA
Abstract :
Echo state networks (ESNs) are a novel form of recurrent neural networks (RNNs) that provide an efficient and powerful computational model approximating nonlinear dynamical systems. A unique feature of an ESN is that a large number of neurons (the “reservoir”) are used, whose synaptic connections are generated randomly, with only the connections from the reservoir to the output modified by learning. Why a large randomly generated fixed RNN gives such excellent performance in approximating nonlinear systems is still not well understood. In this brief, we apply random matrix theory to examine the properties of random reservoirs in ESNs under different topologies (sparse or fully connected) and connection weights (Bernoulli or Gaussian). We quantify the asymptotic gap between the scaling factor bounds for the necessary and sufficient conditions previously proposed for the echo state property. We then show that the state transition mapping is contractive with high probability when only the necessary condition is satisfied, which corroborates and thus analytically explains the observation that in practice one obtains echo states when the spectral radius of the reservoir weight matrix is smaller than 1.
Keywords :
matrix algebra; modelling; nonlinear dynamical systems; recurrent neural nets; echo state networks; nonlinear dynamical systems; nonlinear system modeling; random matrix theory; recurrent neural networks; reservoir weight matrix; scaling factor bounds; state transition mapping; synaptic connections; Eigenvalues and eigenfunctions; Learning systems; Network topology; Neurons; Recurrent neural networks; Reservoirs; Sparse matrices; Circular law; concentration of measure; echo state networks; echo state property; random matrix theory; recurrent neural networks;
Journal_Title :
Neural Networks and Learning Systems, IEEE Transactions on
DOI :
10.1109/TNNLS.2011.2178562