DocumentCode
2324303
Title
Evolving ARTMAP neural networks using Multi-Objective Particle Swarm Optimization
Author
Granger, Eric ; Prieur, Donavan ; Connolly, Jean-François
Author_Institution
Lab. d´´imagerie, de Vision et d´´Intell. Artificielle (LIVIA), Ecole de Technol. Super., Montreal, QC, Canada
fYear
2010
fDate
18-23 July 2010
Firstpage
1
Lastpage
8
Abstract
In this paper, a supervised learning strategy based on a Multi-Objective Particle Swarm Optimization (MOPSO) is introduced for ARTMAP neural networks. It is based on the concept of neural network evolution in that particles of a MOPSO swarm (i.e., network solutions) seek to determine user-defined parameters and network (weights and architecture) such that generalisation error and network resources are minimized. The performance of this strategy has been assessed with fuzzy ARTMAP using synthetic and real-world data for video-based face classification. Simulation results indicate that when the MOPSO strategy is used to train fuzzy ARTMAP, it produces a significantly lower classification error than when trained using standard hyper-parameter settings. Furthermore, the non-dominated MOPSO solutions represent a better compromise between error and resource allocation than mono-objective PSO-based strategies that minimizes only classification error. Overall, results obtained with the MOPSO strategy reveal the importance of optimizing parameters and network for each problem, where both error and resources are minimized during fitness evaluation.
Keywords
ART neural nets; face recognition; fuzzy neural nets; image classification; learning (artificial intelligence); particle swarm optimisation; video signal processing; ARTMAP neural network; MOPSO swarm; classification error; fitness evaluation; fuzzy ARTMAP; generalisation error; multiobjective particle swarm optimization; network resource; neural network evolution; parameter optimization; resource allocation; supervised learning; user-defined parameter; video-based face classification; Artificial neural networks; Convergence; Hypercubes; Lead; Optimization; Supervised learning; Training;
fLanguage
English
Publisher
ieee
Conference_Titel
Evolutionary Computation (CEC), 2010 IEEE Congress on
Conference_Location
Barcelona
Print_ISBN
978-1-4244-6909-3
Type
conf
DOI
10.1109/CEC.2010.5585953
Filename
5585953
Link To Document