Title :
Integrating detailed information into a language model
Author :
Zhang, Ruiqiang ; Black, Ezra ; Finch, Andrew ; Sagisaka, Yoshinori
Author_Institution :
ATR Interpreting Telephony Res. Labs., Kyoto, Japan
Abstract :
Applying natural language processing technique to language modeling is a key problem in speech recognition. This paper describes a maximum entropy-based approach to language modeling in which both words together with syntactic and semantic tags in the long history are used as a basis for complex linguistic questions. These questions are integrated with a standard trigram language model or a standard trigram language model combined with long history word triggers and the resulting language model is used to rescore the N-best hypotheses output of the ATRSPREC speech recognition system. The technique removed 24% of the correctable error of the recognition system
Keywords :
linguistics; maximum entropy methods; natural languages; speech recognition; ATRSPREC speech recognition system; N-best hypotheses output; complex linguistic questions; correctable error; language model; long history word triggers; maximum entropy-based approach; natural language processing technique; semantic tags; speech recognition; standard trigram language model; syntactic tags; Error analysis; Error correction; History; Information resources; Laboratories; Natural language processing; Natural languages; Predictive models; Speech recognition; Vocabulary;
Conference_Titel :
Acoustics, Speech, and Signal Processing, 2000. ICASSP '00. Proceedings. 2000 IEEE International Conference on
Conference_Location :
Istanbul
Print_ISBN :
0-7803-6293-4
DOI :
10.1109/ICASSP.2000.861995