DocumentCode :
671544
Title :
A study of transformation-invariances of deep belief networks
Author :
Zheng Shou ; Yuhao Zhang ; Cai, H.J.
Author_Institution :
Int. Sch. of Software, Wuhan Univ., Wuhan, China
fYear :
2013
fDate :
4-9 Aug. 2013
Firstpage :
1
Lastpage :
8
Abstract :
In order to learn transformation-invariant features, several effective deep architectures like hierarchical feature learning and variant Deep Belief Networks (DBN) have been proposed. Considering the complexity of those variants, people are interested in whether DBN itself has transformation-invariances. First of all, we use original DBN to test original data. Almost same error rates will be achieved, if we change weights in the bottom interlayer according to transformations occurred in testing data. It implies that weights in the bottom interlayer can store the knowledge to handle transformations such as rotation, shifting, and scaling. Along with the continuous learning ability and good storage of DBN, we present our Weight-Transformed Training Algorithm (WTTA) without augmenting other layers, units or filters to original DBN. Based upon original training method, WTTA is aiming at transforming weights and is still unsupervised. For MNIST handwritten digits recognizing experiments, we adopted 784-100-100-100 DBN to compare the differences of recognizing ability in weights-transformed ranges. Most error rates generated by WTTA were below 25% while most rates generated by original training algorithm exceeded 25%. Then we also did an experiment on part of MIT-CBCL face database, with varying illumination, and the best testing accuracy can be achieved is 87.5%. Besides, similar results can be achieved by datasets covering all kinds of transformations, but WTTA only needs original training data and transform weights after each training loop. Consequently, we can mine inherent transformation-invariances of DBN by WTTA, and DBN itself can recognize transformed data at satisfying error rates without inserting other components.
Keywords :
belief networks; feature extraction; handwritten character recognition; lighting; neural nets; object recognition; probability; unsupervised learning; DBN; MIT-CBCL face database; MNIST handwritten digits recognizing experiments; WTTA; deep belief networks; hierarchical feature learning; illumination variation; multilayer probabilistic generative model; robust feature extraction models; rotation transformation; scaling transformation; shifting transformation; training loop; transformation-invariances; transformation-invariant features; unsupervised learning algorithm; weight-transformed training algorithm; Error analysis; Feature extraction; Neurons; Pattern recognition; Training; Transforms;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks (IJCNN), The 2013 International Joint Conference on
Conference_Location :
Dallas, TX
ISSN :
2161-4393
Print_ISBN :
978-1-4673-6128-6
Type :
conf
DOI :
10.1109/IJCNN.2013.6706884
Filename :
6706884
Link To Document :
بازگشت