• DocumentCode
    3132328
  • Title

    Context dependent recurrent neural network language model

  • Author

    Mikolov, T. ; Zweig, Geoffrey

  • Author_Institution
    Brno Univ. of Technol., Brno, Czech Republic
  • fYear
    2012
  • fDate
    2-5 Dec. 2012
  • Firstpage
    234
  • Lastpage
    239
  • Abstract
    Recurrent neural network language models (RNNLMs) have recently demonstrated state-of-the-art performance across a variety of tasks. In this paper, we improve their performance by providing a contextual real-valued input vector in association with each word. This vector is used to convey contextual information about the sentence being modeled. By performing Latent Dirichlet Allocation using a block of preceding text, we achieve a topic-conditioned RNNLM. This approach has the key advantage of avoiding the data fragmentation associated with building multiple topic models on different data subsets. We report perplexity results on the Penn Treebank data, where we achieve a new state-of-the-art. We further apply the model to the Wall Street Journal speech recognition task, where we observe improvements in word-error-rate.
  • Keywords
    natural language processing; recurrent neural nets; speech recognition; text analysis; word processing; Penn Treebank data; Wall Street Journal speech recognition task; context dependent recurrent neural network language model; contextual real-valued input vector; data fragmentation; latent Dirichlet allocation; multiple topic models; perplexity; sentence modelling; text block; topic-conditioned RNNLM; word-error-rate improvement; Computational modeling; Context; Context modeling; Data models; Neurons; Recurrent neural networks; Vectors; Language Modeling; Latent Dirichlet Allocation; Recurrent Neural Network; Topic Models;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Spoken Language Technology Workshop (SLT), 2012 IEEE
  • Conference_Location
    Miami, FL
  • Print_ISBN
    978-1-4673-5125-6
  • Electronic_ISBN
    978-1-4673-5124-9
  • Type

    conf

  • DOI
    10.1109/SLT.2012.6424228
  • Filename
    6424228