DocumentCode
351015
Title
Learning to predict a context-free language: analysis of dynamics in recurrent hidden units
Author
Bodén, Mikael ; Wiles, Janet ; Tonkes, Bradley ; Blair, Alan
Author_Institution
Dept. of Comput. Sci. & Electr. Eng., Queensland Univ., Qld., Australia
Volume
1
fYear
1999
fDate
1999
Firstpage
359
Abstract
Previous results regarding the language anbn suggest that while it is possible for a small recurrent neural network to process context-free languages, learning them is difficult. This paper considers the reasons underlying this difficulty by considering the relationship between the dynamics of the network and weight-space. We show that the dynamics required for the solution lie in a region of weight-space close to a bifurcation point where small changes in weights may result in radically different network behaviour. Furthermore, we show that the error gradient information in this region is highly irregular. We conclude that any gradient-based learning method will experience difficulty in learning the language due to the nature of the space, and that a more promising approach to improving learning performance may be to make weight changes in a non-independent manner
Keywords
recurrent neural nets; context-free language; dynamics; error gradient; learning method; recurrent hidden units; recurrent neural network; weight-space;
fLanguage
English
Publisher
iet
Conference_Titel
Artificial Neural Networks, 1999. ICANN 99. Ninth International Conference on (Conf. Publ. No. 470)
Conference_Location
Edinburgh
ISSN
0537-9989
Print_ISBN
0-85296-721-7
Type
conf
DOI
10.1049/cp:19991135
Filename
819747
Link To Document