• DocumentCode
    353656
  • Title

    Integrating detailed information into a language model

  • Author

    Zhang, Ruiqiang ; Black, Ezra ; Finch, Andrew ; Sagisaka, Yoshinori

  • Author_Institution
    ATR Interpreting Telephony Res. Labs., Kyoto, Japan
  • Volume
    3
  • fYear
    2000
  • fDate
    2000
  • Firstpage
    1595
  • Abstract
    Applying natural language processing technique to language modeling is a key problem in speech recognition. This paper describes a maximum entropy-based approach to language modeling in which both words together with syntactic and semantic tags in the long history are used as a basis for complex linguistic questions. These questions are integrated with a standard trigram language model or a standard trigram language model combined with long history word triggers and the resulting language model is used to rescore the N-best hypotheses output of the ATRSPREC speech recognition system. The technique removed 24% of the correctable error of the recognition system
  • Keywords
    linguistics; maximum entropy methods; natural languages; speech recognition; ATRSPREC speech recognition system; N-best hypotheses output; complex linguistic questions; correctable error; language model; long history word triggers; maximum entropy-based approach; natural language processing technique; semantic tags; speech recognition; standard trigram language model; syntactic tags; Error analysis; Error correction; History; Information resources; Laboratories; Natural language processing; Natural languages; Predictive models; Speech recognition; Vocabulary;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Acoustics, Speech, and Signal Processing, 2000. ICASSP '00. Proceedings. 2000 IEEE International Conference on
  • Conference_Location
    Istanbul
  • ISSN
    1520-6149
  • Print_ISBN
    0-7803-6293-4
  • Type

    conf

  • DOI
    10.1109/ICASSP.2000.861995
  • Filename
    861995