DocumentCode
2701941
Title
Can supervised learning be achieved without explicit error back-propagation?
Author
Brandt, Robert D. ; Lin, Feng
Author_Institution
Beckman Inst. for Adv. Sci. & Technol., Illinois Univ., Urbana, IL, USA
Volume
1
fYear
1996
fDate
3-6 Jun 1996
Firstpage
300
Abstract
We propose a new model for the implementation of supervised learning algorithms for networks of sigmoidal "neurons" which does not require that error feedback be explicitly provided by means of a dedicated feedback network. In this model, a locally-defined environmental gradient which is implicit in the strengths of synapses, their rates of change, and pre- and post-synaptic activity levels is used in the adaptation. This environmental gradient always exists and is generally non-zero, independently of the presence of a supervisor, so long as there is some change in synaptic strength, regardless of the driving force behind the modification, much as a Hebbian gradient always exists at any synapse
Keywords
learning (artificial intelligence); neural nets; Hebbian gradient; error feedback; locally-defined environmental gradient; post-synaptic activity levels; pre-synaptic activity levels; sigmoidal neurons; supervised learning; synaptic strength change; Artificial neural networks; Biological system modeling; Computer errors; Feedforward systems; Neural network hardware; Neurofeedback; Neurons; Neuroscience; Numerical models; Supervised learning;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks, 1996., IEEE International Conference on
Conference_Location
Washington, DC
Print_ISBN
0-7803-3210-5
Type
conf
DOI
10.1109/ICNN.1996.548908
Filename
548908
Link To Document