Title :
Dystal: a self-organizing ANN with pattern independent training time
Author :
Barbour, G. ; Blackwell, K. ; Busse, T. ; Alkon, D. ; Vogl, T.
Author_Institution :
Environ. Res. Inst. of Michigan, Arlington, VA, USA
Abstract :
As the difficulty of problems increase, artificial neural networks (ANNs) that use nonlinear optimization suffer from degraded execution speed, particularly with respect to learning time. Dystal is an ANN which does not suffer this degradation. Dystal is an ANN based on properties of associative learning found in biological neural networks. To verify these theoretical properties of Dystal, the authors implement Dystal on MasPar, a massively parallel machine. They show that the execution time is independent of both the separability of the patterns and the number of output units, and that the training time is linear with the number of patterns in the training data set. That is, the number of iterations through the training set to achieve learning is small and independent of pattern content or training set size
Keywords :
pattern recognition; self-organising feature maps; unsupervised learning; Dystal; MasPar; associative learning; degraded execution speed; learning time; massively parallel machine; nonlinear optimization; pattern independent training time; self-organising artificial neural nets; training data set; Artificial neural networks; Biological neural networks; Biology computing; Computer networks; Degradation; Neurons; Parallel machines; Parallel processing; System testing; Training data;
Conference_Titel :
Neural Networks, 1992. IJCNN., International Joint Conference on
Conference_Location :
Baltimore, MD
Print_ISBN :
0-7803-0559-0
DOI :
10.1109/IJCNN.1992.227217