DocumentCode :
2777893
Title :
Learning sparse representation via a nonlinear shrinkage encoder and a linear sparse decoder
Author :
Ji, Zhengping ; Huang, Wentao ; Brumby, Steven P.
Author_Institution :
Theor. Div., Los Alamos Nat. Lab., Los Alamos, NM, USA
fYear :
2012
fDate :
10-15 June 2012
Firstpage :
1
Lastpage :
8
Abstract :
Learning sparse representations for deep networks has drawn considerable research interest in recent years. In this paper, we present a novel framework to learn sparse representations via a generalized encoder-decoder architecture. The basic idea is to adopt a fast approximation to the iterative sparse coding solution and form an efficient nonlinear encoder to map an input to a sparse representation. A set of basis functions is then learned through the minimization of an energy function consisting of a sparseness prior and linear decoder constraints. Applying a greedy layer-wise learning scheme, this framework can be extended to more layers to learn deep networks. The proposed learning algorithm is also highly efficient as no iterative operations are required, and both batch and on-line learning are supported. Given the sparse representation and basis functions, an optimized decoding procedure is carried out to reconstruct and denoise the input signals. We evaluate our model on natural image patches to develop a dictionary of V1-like Gabor filters, and further show that basis functions in a higher layer (e.g., V2) combine the filters in a lower layer to generate more complex patterns to benefit the high-level tasks. We then use the sparse representations to recognize objects in two benchmark data sets (i.e., CIFAR-10 and NORB) via a linear SVM classifier, and demonstrate better or comparable recognition performances with respect to state-of-art algorithms. The image reconstruction of MNIST images and the restoration of corrupted versions are presented at the end.
Keywords :
Gabor filters; decoding; image coding; image representation; image restoration; iterative methods; learning (artificial intelligence); minimisation; pattern classification; signal denoising; support vector machines; MNIST images; V1-like Gabor filters; basis functions; batch learning; benchmark data sets; complex patterns; corrupted versions restoration; deep networks; dictionary; energy function; generalized encoder-decoder architecture; greedy layer-wise learning scheme; image reconstruction; iterative operations; iterative sparse coding solution; learning algorithm; learning sparse representation; linear SVM classifier; linear sparse decoder; minimization; natural image patches; nonlinear encoder; nonlinear shrinkage encoder; online learning; optimized decoding procedure; recognition performances; research interest; signal denoising; signal reconstruction; sparse representations; sparseness linear decoder constraints; sparseness prior decoder constraints; state-of-art algorithms; Decoding; Dictionaries; Encoding; Image reconstruction; Iterative decoding; Matrix decomposition; Vectors;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks (IJCNN), The 2012 International Joint Conference on
Conference_Location :
Brisbane, QLD
ISSN :
2161-4393
Print_ISBN :
978-1-4673-1488-6
Electronic_ISBN :
2161-4393
Type :
conf
DOI :
10.1109/IJCNN.2012.6252810
Filename :
6252810
Link To Document :
بازگشت