DocumentCode
730195
Title
Classifying phonological categories in imagined and articulated speech
Author
Shunan Zhao ; Rudzicz, Frank
Author_Institution
Dept. of Comput. Sci., Univ. of Toronto, Toronto, ON, Canada
fYear
2015
fDate
19-24 April 2015
Firstpage
992
Lastpage
996
Abstract
This paper presents a new dataset combining 3 modalities (EEG, facial, and audio) during imagined and vocalized phonemic and single-word prompts. We pre-process the EEG data, compute features for all 3 modalities, and perform binary classification of phonological categories using a combination of these modalities. For example, a deep-belief network obtains accuracies over 90% on identifying consonants, which is significantly more accurate than two baseline support vector machines. We also classify between the different states (resting, stimuli, active thinking) of the recording, achieving accuracies of 95%. These data may be used to learn multimodal relationships, and to develop silent-speech and brain-computer interfaces.
Keywords
brain-computer interfaces; electroencephalography; feature extraction; medical signal processing; signal classification; speech processing; support vector machines; EEG dataset; audio dataset; binary classification; brain-computer interfaces; facial dataset; phonological categories; silent-speech; support vector machines; Electroencephalography; Presses; Tin; Phonological categories; deep-belief networks; electroencephalography; speech articulation;
fLanguage
English
Publisher
ieee
Conference_Titel
Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on
Conference_Location
South Brisbane, QLD
Type
conf
DOI
10.1109/ICASSP.2015.7178118
Filename
7178118
Link To Document