• DocumentCode
    730195
  • Title

    Classifying phonological categories in imagined and articulated speech

  • Author

    Shunan Zhao ; Rudzicz, Frank

  • Author_Institution
    Dept. of Comput. Sci., Univ. of Toronto, Toronto, ON, Canada
  • fYear
    2015
  • fDate
    19-24 April 2015
  • Firstpage
    992
  • Lastpage
    996
  • Abstract
    This paper presents a new dataset combining 3 modalities (EEG, facial, and audio) during imagined and vocalized phonemic and single-word prompts. We pre-process the EEG data, compute features for all 3 modalities, and perform binary classification of phonological categories using a combination of these modalities. For example, a deep-belief network obtains accuracies over 90% on identifying consonants, which is significantly more accurate than two baseline support vector machines. We also classify between the different states (resting, stimuli, active thinking) of the recording, achieving accuracies of 95%. These data may be used to learn multimodal relationships, and to develop silent-speech and brain-computer interfaces.
  • Keywords
    brain-computer interfaces; electroencephalography; feature extraction; medical signal processing; signal classification; speech processing; support vector machines; EEG dataset; audio dataset; binary classification; brain-computer interfaces; facial dataset; phonological categories; silent-speech; support vector machines; Electroencephalography; Presses; Tin; Phonological categories; deep-belief networks; electroencephalography; speech articulation;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on
  • Conference_Location
    South Brisbane, QLD
  • Type

    conf

  • DOI
    10.1109/ICASSP.2015.7178118
  • Filename
    7178118