DocumentCode :
330301
Title :
Neural networks: life after training
Author :
Salerno, John J.
Author_Institution :
Air Force Res. Lab., Rome, NY, USA
Volume :
2
fYear :
1998
fDate :
11-14 Oct 1998
Firstpage :
1680
Abstract :
There has been much work done in the use of neural networks to model an existing problem, but little has been done to address what happens after training has been completed and the model must continue to learn new information. How well does the model work on information that it has not seen before? How does it adapt to new information? In this paper we address these issues, beginning our discussion with a neural model that has been trained on parsing simple natural language phrases and how well the model can generalize. Based on these results we then investigate two techniques which attempt to allow the model to “grow” or learn information that it has never before seen
Keywords :
generalisation (artificial intelligence); learning (artificial intelligence); neural nets; generalisation; information learning; natural language; neural networks; update policy; Animals; Backpropagation; Concatenated codes; Data preprocessing; Feeds; Humans; Instruments; Laboratories; Natural languages; Neural networks;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Systems, Man, and Cybernetics, 1998. 1998 IEEE International Conference on
Conference_Location :
San Diego, CA
ISSN :
1062-922X
Print_ISBN :
0-7803-4778-1
Type :
conf
DOI :
10.1109/ICSMC.1998.728135
Filename :
728135
Link To Document :
بازگشت