DocumentCode :
3097348
Title :
Learning with nonstatic paradigms in neural networks
Author :
Köhle, Monika ; Schönbauer, Franz
Author_Institution :
Tech. Univ. of Vienna, Austria
fYear :
1989
fDate :
10-12 Apr 1989
Firstpage :
72
Lastpage :
75
Abstract :
Schemata, concepts, or any kind of knowledge contained in a neural network is usually represented in the interconnections of the constituent units. Learning in neural networks is regarded as modifying the strength of these interconnections. The authors define higher structured learning as not only modifying the weights but also the overall topology of the net, which implies that knowledge in a neural network is also represented in the architecture of the net. In the static paradigm, the definition of a neural network is similar to a variable declaration in a block-oriented language. The overall topology of the network is defined before learning takes place and is not altered during the learning phase. During learning only the values of the weights change. In a nonstatic paradigm units can be created at any time and can be connected arbitrarily with other units of the net. The authors demonstrate that using a nonstatic approach can improve learning
Keywords :
learning systems; neural nets; architecture; block-oriented language; concepts; constituent units; higher structured learning; interconnections; knowledge; neural networks; nonstatic paradigms; schemata; static paradigm; topology; variable declaration; weights; Feature extraction; Feeds; Intelligent networks; Network topology; Neural networks; Pattern recognition; Performance evaluation; Supervised learning; Testing;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Industrial Applications of Machine Intelligence and Vision, 1989., International Workshop on
Conference_Location :
Tokyo
Type :
conf
DOI :
10.1109/MIV.1989.40525
Filename :
40525
Link To Document :
بازگشت