DocumentCode :
1681761
Title :
Extended theory refinement in knowledge-based neural networks
Author :
Garcez, Artur S d Avila
Author_Institution :
Dept. of Comput., Imperial Coll. of Sci., Technol. & Med., London, UK
Volume :
3
fYear :
2002
fDate :
6/24/1905 12:00:00 AM
Firstpage :
2905
Lastpage :
2910
Abstract :
This paper shows that single hidden layer networks with semi-linear activation function compute the answer set semantics of extended logic programs. As a result, incomplete (nonmonotonic) theories, presented as extended logic programs, i.e., possibly containing both classical and default negations, may be refined through inductive learning in knowledge-based neural networks
Keywords :
feedforward neural nets; knowledge based systems; learning (artificial intelligence); logic programming; transfer functions; default negation; extended logic programming; feedforward neural networks; hybrid architectures; inductive learning; knowledge-based neural networks; semilinear activation function; single hidden layer networks; translation algorithm; Artificial neural networks; Computer networks; Concurrent computing; Educational institutions; Hybrid power systems; Intelligent networks; Logic programming; Machine learning; Neural networks; Refining;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2002. IJCNN '02. Proceedings of the 2002 International Joint Conference on
Conference_Location :
Honolulu, HI
ISSN :
1098-7576
Print_ISBN :
0-7803-7278-6
Type :
conf
DOI :
10.1109/IJCNN.2002.1007610
Filename :
1007610
Link To Document :
بازگشت