• DocumentCode
    3607037
  • Title

    Exponential Language Modeling Using Morphological Features and Multi-Task Learning

  • Author

    Hao Fang ; Ostendorf, Mari ; Baumann, Peter ; Pierrehumbert, Janet

  • Author_Institution
    Dept. of Electr. Eng., Univ. of Washington, Seattle, WA, USA
  • Volume
    23
  • Issue
    12
  • fYear
    2015
  • Firstpage
    2410
  • Lastpage
    2421
  • Abstract
    For languages with fast vocabulary growth and limited resources, data sparsity leads to challenges in training a language model. One strategy for addressing this problem is to leverage morphological structure as features in the model. This paper explores different uses of unsupervised morphological features in both the history and prediction space for three word-based exponential models (maximum entropy, logbilinear, and recurrent neural net (RNN)). Multi-task training is introduced as a regularizing mechanism to improve performance in the continuous-space approaches. The models are compared to non-parametric baselines. From using the RNN with morphological features and multi-task learning, experiments with conversational speech from four languages show we can obtain consistent gains of 7-11% in perplexity reduction in a limited-resource scenario (10 hrs speech), and 12-18% when the training size is increased ( 80 hrs ). Results are mixed for all other approaches, compared to a modified Kneser-Ney baseline, but morphology is useful in continuous-space models compared to their word-only baseline. Multi-task learning improves both continuous-space models.
  • Keywords
    learning (artificial intelligence); maximum entropy methods; natural language processing; recurrent neural nets; speech processing; RNN; continuous-space approach; continuous-space models; conversational speech; data sparsity; history space; language model training; limited-resource scenario; logbilinear; maximum entropy; morphological features; morphological structure leveraging; multitask learning; multitask training; performance improvement; perplexity reduction; prediction space; recurrent neural net; regularizing mechanism; training size; unsupervised morphological features; vocabulary growth; word-based exponential language modeling; Context modeling; Data models; Languages; Morphology; Neural networks; Predictive models; Recurrent neural networks; Speech processing; Vocabulary; Language model; limited resources; morphology; neural network;
  • fLanguage
    English
  • Journal_Title
    Audio, Speech, and Language Processing, IEEE/ACM Transactions on
  • Publisher
    ieee
  • ISSN
    2329-9290
  • Type

    jour

  • DOI
    10.1109/TASLP.2015.2482118
  • Filename
    7277009