Title :
Discriminative auditory features for robust speech recognition
Author :
Mak, Brian ; Tam, Yik-Cheung ; Li, Qi
Author_Institution :
Hong Kong University of Science and Technology, Department of Computer Science, Clear Water Bay, Hong Kong
Abstract :
Recently, Li et al. proposed a new auditory feature for robust speech recognition in noise environments. The new feature was derived by mimicking closely the function of human auditory process. Several filters were used to model the outer ear, middle ear, and cochlea, and the initial filter parameters and shapes were obtained from crude psychoacoustics results, experience, or experiments. Although one may adjust the feature parameters by hand to get better performance, the resulting feature parameters still may not be optimal in the sense of minimal recognition errors, especially for different tasks. To further improve the auditory feature, in this paper we apply discriminative training to optimize the auditory feature parameters with some guidance from psychoacoustic evidence but otherwise in a data-driven approach so as to minimize the recognition errors. One significant contribution over similar efforts in the past, such as discriminative feature extraction, is that we make no assumption on the parametric form of the auditory filters. Instead, we only require the filters to be smooth and triangular-like as suggested by psychoacoustics research. Our approach is evaluated on the Aurora database and achieves a word error reduction of 19.2%.
Keywords :
Feature extraction; Hidden Markov models; Indexes; Information filters; Training;
Conference_Titel :
Acoustics, Speech, and Signal Processing (ICASSP), 2002 IEEE International Conference on
Print_ISBN :
0-7803-7402-9
DOI :
10.1109/ICASSP.2002.5743734