• DocumentCode
    1020901
  • Title

    Multilayer perceptrons as labelers for hidden Markov models

  • Author

    Le Cerf, P. ; Ma, Weiye ; Van Compernolle, D.

  • Author_Institution
    ESAT, Katholieke Univ., Leuven, Heverlee, Belgium
  • Volume
    2
  • Issue
    1
  • fYear
    1994
  • Firstpage
    185
  • Lastpage
    193
  • Abstract
    A novel combination of multilayer perceptrons (MLPs) and hidden Markov models (HMMs) is presented. Instead of using MLPs as probability generators for HMMs, the authors propose to use MLPs as labelers for discrete parameter HMM´s. Compared with the probabilistic interpretation of MLPs, this gives them the advantage of flexibility in system design (e.g., the use of word models instead of phonetic models while using the same MLPs). Moreover, since they do not need to reach a global minimum, they can do with MLPs with fewer hidden nodes, which can be trained faster. In addition, they do not need to retrain the MLPs with segmentations generated by a Viterbi alignment. Compared with Euclidean labeling, their method has the advantages of needing fewer HMM parameters per state and obtaining a higher recognition accuracy. Several improvements of the baseline MLP labeling are investigated. When using one MLP, the best results are obtained when giving the labels a fuzzy interpretation. It is also possible to use parallel MLPs where each is based on a different parameter set (e.g., basic parameters, their time derivatives, and their second-order time derivatives). This strategy increases the recognition results considerably. A final improvement is the training of MLPs for subphoneme classification.
  • Keywords
    feedforward neural nets; hidden Markov models; learning (artificial intelligence); probability; Euclidean labeling; HMM parameters; MLP; Viterbi alignment; baseline MLP labeling; discrete parameter labelers; fuzzy interpretation; hidden Markov models; hidden nodes; multilayer perceptrons; probabilistic estimators; recognition accuracy; second-order time derivatives; speech recognition; subphoneme classification; system design; training; word models; Artificial neural networks; Brain modeling; Hidden Markov models; Labeling; Multilayer perceptrons; Neural networks; Pattern recognition; Recurrent neural networks; Speech recognition; Viterbi algorithm;
  • fLanguage
    English
  • Journal_Title
    Speech and Audio Processing, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    1063-6676
  • Type

    jour

  • DOI
    10.1109/89.260361
  • Filename
    260361