DocumentCode :
3748606
Title :
Deep Fried Convnets
Author :
Zichao Yang;Marcin Moczulski;Misha Denil;Nando de Freitas;Alex Smola;Le Song;Ziyu Wang
fYear :
2015
Firstpage :
1476
Lastpage :
1483
Abstract :
The fully-connected layers of deep convolutional neural networks typically contain over 90% of the network parameters. Reducing the number of parameters while preserving predictive performance is critically important for training big models in distributed systems and for deployment in embedded devices. In this paper, we introduce a novel Adaptive Fastfood transform to reparameterize the matrix-vector multiplication of fully connected layers. Reparameterizing a fully connected layer with d inputs and n outputs with the Adaptive Fastfood transform reduces the storage and computational costs costs from O(nd) to O(n) and O(n log d) respectively. Using the Adaptive Fastfood transform in convolutional networks results in what we call a deep fried convnet. These convnets are end-to-end trainable, and enable us to attain substantial reductions in the number of parameters without affecting prediction accuracy on the MNIST and ImageNet datasets.
Keywords :
"Transforms","Kernel","Neural networks","Sparse matrices","Training","Adaptive systems","Computational efficiency"
Publisher :
ieee
Conference_Titel :
Computer Vision (ICCV), 2015 IEEE International Conference on
Electronic_ISBN :
2380-7504
Type :
conf
DOI :
10.1109/ICCV.2015.173
Filename :
7410530
Link To Document :
بازگشت