Title :
Backpropagation algorithm for logic oriented neural networks
Author :
Kamio, Takeshi ; Tanaka, Shinichiro ; Morisue, Mititada
Author_Institution :
Hiroshima City Univ., Hiroshima, Japan
Abstract :
Multilayer feedforward neural network (MFNN) trained by the backpropagation (BP) algorithm is one of the most significant models in artificial neural networks. Although they have been implemented as analog, mixed analog-digital and fully digital VLSI circuits, it is still difficult to realize their hardware implementation with BP learning function. This paper describes the BP algorithm for the logic oriented neural network (LOGO-NN) which we have proposed as a kind of MFNN with quantized weights and multilevel threshold neurons. Since both weights and neuron outputs are quantized to integer values in LOGO-NNs, it is expected that LOGO-NNs with BP learning can be more effectively implemented than the common MFNNs. Finally, it is shown by simulations that the proposed BP algorithm has good performance for LOGO-NNs
Keywords :
backpropagation; convergence; feedforward neural nets; formal logic; pattern recognition; backpropagation; convergence; feedforward neural network; learning function; logic oriented neural networks; multilevel threshold neurons; pattern recognition; quantized weights; Analog-digital conversion; Artificial neural networks; Backpropagation algorithms; Circuits; Feedforward neural networks; Logic; Multi-layer neural network; Neural networks; Neurons; Very large scale integration;
Conference_Titel :
Neural Networks, 2000. IJCNN 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on
Conference_Location :
Como
Print_ISBN :
0-7695-0619-4
DOI :
10.1109/IJCNN.2000.857885