DocumentCode
353656
Title
Integrating detailed information into a language model
Author
Zhang, Ruiqiang ; Black, Ezra ; Finch, Andrew ; Sagisaka, Yoshinori
Author_Institution
ATR Interpreting Telephony Res. Labs., Kyoto, Japan
Volume
3
fYear
2000
fDate
2000
Firstpage
1595
Abstract
Applying natural language processing technique to language modeling is a key problem in speech recognition. This paper describes a maximum entropy-based approach to language modeling in which both words together with syntactic and semantic tags in the long history are used as a basis for complex linguistic questions. These questions are integrated with a standard trigram language model or a standard trigram language model combined with long history word triggers and the resulting language model is used to rescore the N-best hypotheses output of the ATRSPREC speech recognition system. The technique removed 24% of the correctable error of the recognition system
Keywords
linguistics; maximum entropy methods; natural languages; speech recognition; ATRSPREC speech recognition system; N-best hypotheses output; complex linguistic questions; correctable error; language model; long history word triggers; maximum entropy-based approach; natural language processing technique; semantic tags; speech recognition; standard trigram language model; syntactic tags; Error analysis; Error correction; History; Information resources; Laboratories; Natural language processing; Natural languages; Predictive models; Speech recognition; Vocabulary;
fLanguage
English
Publisher
ieee
Conference_Titel
Acoustics, Speech, and Signal Processing, 2000. ICASSP '00. Proceedings. 2000 IEEE International Conference on
Conference_Location
Istanbul
ISSN
1520-6149
Print_ISBN
0-7803-6293-4
Type
conf
DOI
10.1109/ICASSP.2000.861995
Filename
861995
Link To Document