Title :
Push-pull separability objective for supervised layer-wise training of neural networks
Author :
Szymanski, Lech ; McCane, Brendan
Author_Institution :
Dept. of Comput. Sci., Univ. of Otago, Dunedin, New Zealand
Abstract :
Deep architecture neural networks have been shown to generalise well for many classification problems, however, outside the empirical evidence, it is not entirely clear how deep representation benefits these problems. This paper proposes a supervised cost function for an individual layer in a deep architecture classifier that improves data separability. From this measure, a training algorithm for a multi-layer neural network is developed and evaluated against backpropagation and deep belief net learning. The results confirm that the proposed supervised training objective leads to appropriate internal representation with respect to the classification task, especially for datasets where unsupervised pre-conditioning is not effective. Separability of the hidden layers offers new directions and insights in the quest to illuminate the black box model of deep architectures.
Keywords :
backpropagation; learning (artificial intelligence); neural nets; pattern classification; backpropagation; black box model; classification problems; data separability; deep architecture classifier; deep architecture neural networks; deep belief net learning; deep representation; multilayer neural network; push-pull separability objective; supervised cost function; supervised layer-wise neural network training; unsupervised preconditioning; Backpropagation; Computer architecture; Equations; Neural networks; Neurons; Training; Vectors;
Conference_Titel :
Neural Networks (IJCNN), The 2012 International Joint Conference on
Conference_Location :
Brisbane, QLD
Print_ISBN :
978-1-4673-1488-6
Electronic_ISBN :
2161-4393
DOI :
10.1109/IJCNN.2012.6252366