DocumentCode :
3208128
Title :
Global optimization methods for designing and training neural networks
Author :
Yamazaki, Akio ; Ludermir, Teresa B. ; De Souto, Marcílio C P
Author_Institution :
Center of Informatics, Univ. Fed. de Pernambuco, Recife, Brazil
fYear :
2002
fDate :
2002
Firstpage :
136
Lastpage :
141
Abstract :
This paper shows results of two approaches for the optimization of neural networks: one uses simulated annealing for optimizing both architectures and weights combined with backpropagation for fine tuning, while the other uses tabu search for the same purpose. Both approaches generate networks with good generalization performance (mean classification error of 1.68% for simulated annealing and 0.64% for tabu search) and low complexity (mean number of connections of 11.15 out of 36 for simulated annealing and 11.62 out of 36 for tabu search) for an odor recognition task in an artificial nose.
Keywords :
backpropagation; generalisation (artificial intelligence); neural nets; pattern classification; search problems; simulated annealing; artificial nose; backpropagation; generalization; neural networks; odor recognition; optimization; simulated annealing; tabu search; Artificial neural networks; Backpropagation; Costs; Design methodology; Informatics; Information processing; Neural networks; Nose; Optimization methods; Simulated annealing;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2002. SBRN 2002. Proceedings. VII Brazilian Symposium on
Print_ISBN :
0-7695-1709-9
Type :
conf
DOI :
10.1109/SBRN.2002.1181455
Filename :
1181455
Link To Document :
بازگشت