DocumentCode
1099031
Title
Voiced/Unvoiced/Mixed excitation classification of speech
Author
Siegel, Leah J. ; Bessey, Alan C.
Author_Institution
Princeton University, West Lafayette, NJ
Volume
30
Issue
3
fYear
1982
fDate
6/1/1982 12:00:00 AM
Firstpage
451
Lastpage
460
Abstract
Methods for performing voiced/unvoiced/mixed excitation classification of speech are explored. The decision-making process is viewed as a pattern recognition problem. Three aspects of the task are examined: classifier type, decision structure, and feature selection. A variety of different approaches are compared. A classifier is obtained which, in limited tests, achieves 95 percent classification accuracy on speaker dependent tests (with 82.7 percent correct identification of mixed excitation frames), and 94 percent accuracy on speaker independent tests (with 77.6 percent correct identification of mixed excitation frames). The classifier uses a binary decision tree structure, in which a speech segment is first classified as predominantly voiced or predominantly unvoiced, then tested to determine if the excitation for the segment is mixed or not. Each decision is made using a Bayes classifier. The feature selection procedure identified a set of 14 features to make the voiced/unvoiced/mixed excitation classification.
Keywords
Classification tree analysis; Filters; Helium; Pattern recognition; Speech analysis; Speech processing; Speech recognition; Speech synthesis; Testing; Tree data structures;
fLanguage
English
Journal_Title
Acoustics, Speech and Signal Processing, IEEE Transactions on
Publisher
ieee
ISSN
0096-3518
Type
jour
DOI
10.1109/TASSP.1982.1163910
Filename
1163910
Link To Document