Title :
Convergence models for Rosenblatt´s perceptron learning algorithm
Author :
Diggavi, Suhas N. ; Shynk, John J. ; Bershad, Neil J.
Author_Institution :
Dept. of Electr. & Comput. Eng., California Univ., Santa Barbara, CA, USA
fDate :
7/1/1995 12:00:00 AM
Abstract :
Presents a stochastic analysis of the steady-state and transient convergence properties of a single-layer perceptron for fast learning (large step-size, input-power product). The training data are modeled using a system identification formulation with zero-mean Gaussian inputs. The perceptron weights are adjusted by a learning algorithm equivalent to Rosenblatt´s perceptron convergence procedure. It is shown that the convergence points of the algorithm depend on the step size μ and the input signal power (variance) σx2 , and that the algorithm is stable essentially for μ>0. Two coupled nonlinear recursions are derived that accurately model the transient behavior of the algorithm. The authors also examine how these convergence results are affected by noisy perceptron input vectors. Computer simulations are presented to verify the analytical models
Keywords :
Gaussian processes; convergence; learning (artificial intelligence); perceptrons; signal processing; Rosenblatt´s perceptron learning algorithm; analytical models; coupled nonlinear recursions; input signal power; learning algorithm; noisy perceptron input vectors; single-layer perceptron; steady-state convergence properties; step size; stochastic analysis; system identification formulation; training data; transient convergence properties; zero-mean Gaussian inputs; Analytical models; Computer simulation; Convergence; Couplings; Power system modeling; Steady-state; Stochastic processes; System identification; Training data; Transient analysis;
Journal_Title :
Signal Processing, IEEE Transactions on