DocumentCode
1020901
Title
Multilayer perceptrons as labelers for hidden Markov models
Author
Le Cerf, P. ; Ma, Weiye ; Van Compernolle, D.
Author_Institution
ESAT, Katholieke Univ., Leuven, Heverlee, Belgium
Volume
2
Issue
1
fYear
1994
Firstpage
185
Lastpage
193
Abstract
A novel combination of multilayer perceptrons (MLPs) and hidden Markov models (HMMs) is presented. Instead of using MLPs as probability generators for HMMs, the authors propose to use MLPs as labelers for discrete parameter HMM´s. Compared with the probabilistic interpretation of MLPs, this gives them the advantage of flexibility in system design (e.g., the use of word models instead of phonetic models while using the same MLPs). Moreover, since they do not need to reach a global minimum, they can do with MLPs with fewer hidden nodes, which can be trained faster. In addition, they do not need to retrain the MLPs with segmentations generated by a Viterbi alignment. Compared with Euclidean labeling, their method has the advantages of needing fewer HMM parameters per state and obtaining a higher recognition accuracy. Several improvements of the baseline MLP labeling are investigated. When using one MLP, the best results are obtained when giving the labels a fuzzy interpretation. It is also possible to use parallel MLPs where each is based on a different parameter set (e.g., basic parameters, their time derivatives, and their second-order time derivatives). This strategy increases the recognition results considerably. A final improvement is the training of MLPs for subphoneme classification.
Keywords
feedforward neural nets; hidden Markov models; learning (artificial intelligence); probability; Euclidean labeling; HMM parameters; MLP; Viterbi alignment; baseline MLP labeling; discrete parameter labelers; fuzzy interpretation; hidden Markov models; hidden nodes; multilayer perceptrons; probabilistic estimators; recognition accuracy; second-order time derivatives; speech recognition; subphoneme classification; system design; training; word models; Artificial neural networks; Brain modeling; Hidden Markov models; Labeling; Multilayer perceptrons; Neural networks; Pattern recognition; Recurrent neural networks; Speech recognition; Viterbi algorithm;
fLanguage
English
Journal_Title
Speech and Audio Processing, IEEE Transactions on
Publisher
ieee
ISSN
1063-6676
Type
jour
DOI
10.1109/89.260361
Filename
260361
Link To Document