Title :
Improving deep neural networks with multilayer maxout networks
Author :
Weichen Sun ; Fei Su ; Leiquan Wang
Author_Institution :
Sch. of Inf. & Commun. Eng., Beijing Univ. of Posts & Telecommun., Beijing, China
Abstract :
For the purpose of enhancing discriminability of convolutional neural networks (CNNs) and facilitating optimization, a multilayer structured variant of the maxout unit (named Multilayer Maxout Network, MMN) is proposed in this paper. CNNs with maxout units employ linear convolution filters followed by maxout units to abstract representations from less abstract ones. Our model instead applies MMNs as activation functions of CNNs to abstract representations, which inherits advantages of both maxout units and deep neural networks, and is a more general nonlinear function approximator as well. Experimental results show that our proposed model yields better performance on three image classification benchmark datasets (CIFAR-10, CIFAR-100 and MNIST) than some state-of-the-art methods. Furthermore, the influence of MMN in different hidden layers is analyzed, and a trade-off scheme between the accuracy and computing resources is given.
Keywords :
filtering theory; function approximation; image classification; multilayer perceptrons; CNNs; MMN; abstract representations; activation functions; convolutional neural networks; deep neural networks; general nonlinear function approximator; image classification benchmark datasets; linear convolution filters; multilayer maxout networks; multilayer structured variant; Convolution; Educational institutions; Error analysis; Feature extraction; Neural networks; Nonhomogeneous media; Training; Convolutional neural network; Deep learning; Image classification; Maxout; Representation learning;
Conference_Titel :
Visual Communications and Image Processing Conference, 2014 IEEE
DOI :
10.1109/VCIP.2014.7051574