DocumentCode :
296119
Title :
Relaxing backpropagation networks as associative memories
Author :
Peng, Yun ; Zhou, Zonglin ; McClenney, Erik
Author_Institution :
Dept. of Comput. Sci. & Electr. Eng., Maryland Univ., Baltimore, MD, USA
Volume :
4
fYear :
1995
fDate :
Nov/Dec 1995
Firstpage :
1777
Abstract :
One of the most attractive features of associative memories (AM) is their abilities of associative recall, especially recall by incomplete or noisy inputs. However, most existing neural network AM models suffer from their limited storage capacities. In this paper a new model for autoassociative memory based on a novel use of backpropagation (BP) networks is proposed. In this model, recording is done by establishing pattern auto-associations using BP learning, and recall by continuously feeding back the output to the input until the network relaxes into a stable state. The convergence of the recall process is analyzed. The validity of the model is tested by computer experiments. The experiment results show that O(2n) patterns of length n can be correctly stored, and the recall quality with noisy inputs compares very favorably to conventional AM models
Keywords :
backpropagation; content-addressable storage; convergence; neural nets; relaxation theory; associative memories; associative recall; autoassociative memory; continuous feedback; incomplete inputs; limited storage capacities; neural network; noisy inputs; recall process convergence; relaxing backpropagation networks; Active noise reduction; Associative memory; Backpropagation; Computer science; Convergence; Correlation; Feedforward neural networks; Neural networks; Symmetric matrices; Testing;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1995. Proceedings., IEEE International Conference on
Conference_Location :
Perth, WA
Print_ISBN :
0-7803-2768-3
Type :
conf
DOI :
10.1109/ICNN.1995.488890
Filename :
488890
Link To Document :
بازگشت