Title :
Where to apply dropout in recurrent neural networks for handwriting recognition?
Author :
Théodore Bluche;Christopher Kermorvant;Jérôme Louradour
Author_Institution :
LIMSI CNRS, Spoken Language Processing Group, Orsay, France
Abstract :
The dropout technique is a data-driven regularization method for neural networks. It consists in randomly setting some activations from a given hidden layer to zero during training. Repeating the procedure for each training example, it is equivalent to sample a network from an exponential number of architectures that share weights. The goal of dropout is to prevent feature detectors to rely on each other. Dropout has successfully been applied to Deep MLPs and to convolutional neural networks, for various tasks of Speech Recognition and Computer Vision. We recently proposed a way to use dropout in MDLSTM-RNNs for handwritten word and line recognition. In this paper, we show that further improvement can be achieved by implementing dropout differently, more specifically by applying it at better positions relative to the LSTM units.
Keywords :
"Databases","Hidden Markov models","Handwriting recognition","Artificial neural networks","Logic gates","Analytical models"
Conference_Titel :
Document Analysis and Recognition (ICDAR), 2015 13th International Conference on
DOI :
10.1109/ICDAR.2015.7333848