DocumentCode
1807760
Title
A learning method for synthesizing associative memory in neural networks
Author
Kuroe, Yasuaki ; Koashi, Kenshu ; Hashimoto, Naoki ; Mori, Takehiro
Author_Institution
Dept. of Electron. & Inf. Sci., Kyoto Inst. of Technol., Japan
Volume
2
fYear
1999
fDate
36342
Firstpage
798
Abstract
The paper proposes a learning method for synthesizing associative memory in neural networks. The problem is formulated as determining the weights of the synaptic connections of neural networks such that, for any given set of desired memory vectors, each memory vector becomes an asymptotically stable equilibrium point of the network. We introduce a new architecture of neural networks, hybrid recurrent neural networks, in order to enhance the capability of implementing associative memories. An efficient learning method for synthesizing associative memories is proposed. The proposed method assures that all the memory vectors become asymptotically stable equilibrium points with the prescribed degree of stability. Synthesis examples are presented to demonstrate the applicability and performance of the proposed method
Keywords
asymptotic stability; content-addressable storage; learning (artificial intelligence); neural net architecture; recurrent neural nets; associative memory; asymptotically stable equilibrium point; asymptotically stable equilibrium points; hybrid recurrent neural networks; learning method; memory vectors; synaptic connections; Artificial neural networks; Associative memory; Computer networks; Hopfield neural networks; Intelligent networks; Learning systems; Network synthesis; Neural networks; Neurons; Recurrent neural networks;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks, 1999. IJCNN '99. International Joint Conference on
Conference_Location
Washington, DC
ISSN
1098-7576
Print_ISBN
0-7803-5529-6
Type
conf
DOI
10.1109/IJCNN.1999.831052
Filename
831052
Link To Document