DocumentCode :
3292974
Title :
A sequential adder using recurrent networks
Author :
Tsung, Fu-Sheng ; Cottrell, Garrison W.
Author_Institution :
Dept. of Comput. Sci. & Eng., California Univ., San Diego, CA, USA
fYear :
1989
fDate :
0-0 1989
Firstpage :
133
Abstract :
D.E. Rumelhart et al.´s proposal (1986) of how symbolic processing is achieved in PDP (parallel distributed processing) networks is tested by training two types of recurrent networks to learn to add two numbers of arbitrary lengths. A method of combining old and new training sets is developed which enables the network to learn and generalize with very large training sets. Through this model of addition, these networks demonstrated capability to do simple conditional branching, while loops, and sequences, mechanisms essential for a universal computer. Differences between the two types of recurrent networks are discussed, as well as implications for human learning.<>
Keywords :
adders; learning systems; neural nets; parallel architectures; combined subset training; conditional branching; learning; neural networks; parallel distributed processing; recurrent networks; sequential adder; symbolic processing; Adders; Learning systems; Neural networks; Parallel architectures;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1989. IJCNN., International Joint Conference on
Conference_Location :
Washington, DC, USA
Type :
conf
DOI :
10.1109/IJCNN.1989.118690
Filename :
118690
Link To Document :
بازگشت