• DocumentCode
    1099031
  • Title

    Voiced/Unvoiced/Mixed excitation classification of speech

  • Author

    Siegel, Leah J. ; Bessey, Alan C.

  • Author_Institution
    Princeton University, West Lafayette, NJ
  • Volume
    30
  • Issue
    3
  • fYear
    1982
  • fDate
    6/1/1982 12:00:00 AM
  • Firstpage
    451
  • Lastpage
    460
  • Abstract
    Methods for performing voiced/unvoiced/mixed excitation classification of speech are explored. The decision-making process is viewed as a pattern recognition problem. Three aspects of the task are examined: classifier type, decision structure, and feature selection. A variety of different approaches are compared. A classifier is obtained which, in limited tests, achieves 95 percent classification accuracy on speaker dependent tests (with 82.7 percent correct identification of mixed excitation frames), and 94 percent accuracy on speaker independent tests (with 77.6 percent correct identification of mixed excitation frames). The classifier uses a binary decision tree structure, in which a speech segment is first classified as predominantly voiced or predominantly unvoiced, then tested to determine if the excitation for the segment is mixed or not. Each decision is made using a Bayes classifier. The feature selection procedure identified a set of 14 features to make the voiced/unvoiced/mixed excitation classification.
  • Keywords
    Classification tree analysis; Filters; Helium; Pattern recognition; Speech analysis; Speech processing; Speech recognition; Speech synthesis; Testing; Tree data structures;
  • fLanguage
    English
  • Journal_Title
    Acoustics, Speech and Signal Processing, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    0096-3518
  • Type

    jour

  • DOI
    10.1109/TASSP.1982.1163910
  • Filename
    1163910