Title :
Strategies for training large scale neural network language models
Author :
Tomáš Mikolov;Anoop Deoras;Daniel Povey;Lukáš Burget;Jan Černocký
Author_Institution :
Brno University of Technology, Speech@FIT, Czech Republic
Abstract :
We describe how to effectively train neural network based language models on large data sets. Fast convergence during training and better overall performance is observed when the training data are sorted by their relevance. We introduce hash-based implementation of a maximum entropy model, that can be trained as a part of the neural network model. This leads to significant reduction of computational complexity. We achieved around 10% relative reduction of word error rate on English Broadcast News speech recognition task, against large 4-gram model trained on 400M tokens.
Keywords :
"Computational modeling","Training","Artificial neural networks","Data models","Entropy","Computational complexity","Training data"
Conference_Titel :
Automatic Speech Recognition and Understanding (ASRU), 2011 IEEE Workshop on
Print_ISBN :
978-1-4673-0365-1
DOI :
10.1109/ASRU.2011.6163930