DocumentCode
809136
Title
Convergence models for Rosenblatt´s perceptron learning algorithm
Author
Diggavi, Suhas N. ; Shynk, John J. ; Bershad, Neil J.
Author_Institution
Dept. of Electr. & Comput. Eng., California Univ., Santa Barbara, CA, USA
Volume
43
Issue
7
fYear
1995
fDate
7/1/1995 12:00:00 AM
Firstpage
1696
Lastpage
1702
Abstract
Presents a stochastic analysis of the steady-state and transient convergence properties of a single-layer perceptron for fast learning (large step-size, input-power product). The training data are modeled using a system identification formulation with zero-mean Gaussian inputs. The perceptron weights are adjusted by a learning algorithm equivalent to Rosenblatt´s perceptron convergence procedure. It is shown that the convergence points of the algorithm depend on the step size μ and the input signal power (variance) σx2 , and that the algorithm is stable essentially for μ>0. Two coupled nonlinear recursions are derived that accurately model the transient behavior of the algorithm. The authors also examine how these convergence results are affected by noisy perceptron input vectors. Computer simulations are presented to verify the analytical models
Keywords
Gaussian processes; convergence; learning (artificial intelligence); perceptrons; signal processing; Rosenblatt´s perceptron learning algorithm; analytical models; coupled nonlinear recursions; input signal power; learning algorithm; noisy perceptron input vectors; single-layer perceptron; steady-state convergence properties; step size; stochastic analysis; system identification formulation; training data; transient convergence properties; zero-mean Gaussian inputs; Analytical models; Computer simulation; Convergence; Couplings; Power system modeling; Steady-state; Stochastic processes; System identification; Training data; Transient analysis;
fLanguage
English
Journal_Title
Signal Processing, IEEE Transactions on
Publisher
ieee
ISSN
1053-587X
Type
jour
DOI
10.1109/78.398729
Filename
398729
Link To Document