Title :
Recurrent neural network language model with part-of-speech for Mandarin speech recognition
Author :
Caixia Gong ; Xiangang Li ; Xihong Wu
Author_Institution :
Key Lab. of Machine Perception (Minist. of Educ.), Peking Univ., Beijing, China
Abstract :
Recurrent neural network language models (RNNLMs) have been successfully applied in a variety of language processing applications ranging from speech recognition to machine translation. They can fight the curse of dimensionality by learning a distributed representation (word vector). The components of these vectors measure the co-occurrence of the word with context features over a corpus. However, RNNLMs ignore the fact that the meaning of word can vary substantially in different contexts (e.g., for polysemous words). In this paper, we investigate part-of-speech information to address this issue to some extent on the basis of information about the meaning of a word they could provide. Experimental results on Mandarin speech recognition task show that a significant character error reduction of 1.18% absolute (7.72% relative) was obtained when using recurrent neural network language model with part-of-speech.
Keywords :
natural language processing; recurrent neural nets; speech recognition; Mandarin speech recognition task; RNNLM; character error reduction; language processing applications; machine translation; part-of-speech information; recurrent neural network language model; Acoustics; Computational modeling; Dictionaries; Recurrent neural networks; Refining; Speech; Speech recognition; part-of-speech; recurrent neural network language model; speech recognition;
Conference_Titel :
Chinese Spoken Language Processing (ISCSLP), 2014 9th International Symposium on
Conference_Location :
Singapore
DOI :
10.1109/ISCSLP.2014.6936636