DocumentCode
3570652
Title
Improving deep neural networks with multilayer maxout networks
Author
Weichen Sun ; Fei Su ; Leiquan Wang
Author_Institution
Sch. of Inf. & Commun. Eng., Beijing Univ. of Posts & Telecommun., Beijing, China
fYear
2014
Firstpage
334
Lastpage
337
Abstract
For the purpose of enhancing discriminability of convolutional neural networks (CNNs) and facilitating optimization, a multilayer structured variant of the maxout unit (named Multilayer Maxout Network, MMN) is proposed in this paper. CNNs with maxout units employ linear convolution filters followed by maxout units to abstract representations from less abstract ones. Our model instead applies MMNs as activation functions of CNNs to abstract representations, which inherits advantages of both maxout units and deep neural networks, and is a more general nonlinear function approximator as well. Experimental results show that our proposed model yields better performance on three image classification benchmark datasets (CIFAR-10, CIFAR-100 and MNIST) than some state-of-the-art methods. Furthermore, the influence of MMN in different hidden layers is analyzed, and a trade-off scheme between the accuracy and computing resources is given.
Keywords
filtering theory; function approximation; image classification; multilayer perceptrons; CNNs; MMN; abstract representations; activation functions; convolutional neural networks; deep neural networks; general nonlinear function approximator; image classification benchmark datasets; linear convolution filters; multilayer maxout networks; multilayer structured variant; Convolution; Educational institutions; Error analysis; Feature extraction; Neural networks; Nonhomogeneous media; Training; Convolutional neural network; Deep learning; Image classification; Maxout; Representation learning;
fLanguage
English
Publisher
ieee
Conference_Titel
Visual Communications and Image Processing Conference, 2014 IEEE
Type
conf
DOI
10.1109/VCIP.2014.7051574
Filename
7051574
Link To Document