DocumentCode :
730195
Title :
Classifying phonological categories in imagined and articulated speech
Author :
Shunan Zhao ; Rudzicz, Frank
Author_Institution :
Dept. of Comput. Sci., Univ. of Toronto, Toronto, ON, Canada
fYear :
2015
fDate :
19-24 April 2015
Firstpage :
992
Lastpage :
996
Abstract :
This paper presents a new dataset combining 3 modalities (EEG, facial, and audio) during imagined and vocalized phonemic and single-word prompts. We pre-process the EEG data, compute features for all 3 modalities, and perform binary classification of phonological categories using a combination of these modalities. For example, a deep-belief network obtains accuracies over 90% on identifying consonants, which is significantly more accurate than two baseline support vector machines. We also classify between the different states (resting, stimuli, active thinking) of the recording, achieving accuracies of 95%. These data may be used to learn multimodal relationships, and to develop silent-speech and brain-computer interfaces.
Keywords :
brain-computer interfaces; electroencephalography; feature extraction; medical signal processing; signal classification; speech processing; support vector machines; EEG dataset; audio dataset; binary classification; brain-computer interfaces; facial dataset; phonological categories; silent-speech; support vector machines; Electroencephalography; Presses; Tin; Phonological categories; deep-belief networks; electroencephalography; speech articulation;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on
Conference_Location :
South Brisbane, QLD
Type :
conf
DOI :
10.1109/ICASSP.2015.7178118
Filename :
7178118
Link To Document :
بازگشت