Title of article :
A Tsallis’ statistics based neural network model for novel word learning
Author/Authors :
Tarik Hadzibeganovic، نويسنده , , Sergio A. Cannas، نويسنده ,
Issue Information :
روزنامه با شماره پیاپی سال 2009
Abstract :
We invoke the Tsallis entropy formalism, a nonextensive entropy measure, to include some degree of non-locality in a neural network that is used for simulation of novel word learning in adults. A generalization of the gradient descent dynamics, realized via nonextensive cost functions, is used as a learning rule in a simple perceptron. The model is first investigated for general properties, and then tested against the empirical data, gathered from simple memorization experiments involving two populations of linguistically different subjects. Numerical solutions of the model equations corresponded to the measured performance states of human learners. In particular, we found that the memorization tasks were executed with rather small but population-specific amounts of nonextensivity, quantified by the entropic index q. Our findings raise the possibility of using entropic nonextensivity as a means of characterizing the degree of complexity of learning in both natural and artificial systems.
Journal title :
Physica A Statistical Mechanics and its Applications
Journal title :
Physica A Statistical Mechanics and its Applications