• DocumentCode
    2491244
  • Title

    Adaptive context recognition based on audio signal

  • Author

    Zeng, Zhi ; Li, Xin ; Ma, Xiaohong ; Ji, Qiang

  • Author_Institution
    Rensselaer Polytech. Inst., Troy, NY
  • fYear
    2008
  • fDate
    8-11 Dec. 2008
  • Firstpage
    1
  • Lastpage
    4
  • Abstract
    Auditory data provide many contextual cues about the crucial content of environments around. The goal of audio based context recognition is to equip the sensing devices with classification algorithms that can automatically classify the environments into pre-defined classes according to the extracted auditory features. In this paper, we first extract various features from the audio signals. We then perform a feature analysis to identify a feature ensemble to optimally classify different contexts. To achieve an efficient and timely online classification, a coarse-to-fine training scheme is adopted, where for each context three HMMs are trained by feature ensembles of different complexities. During online recognition, we start with coarse HMMs (with fewest numbers of features) and progressively apply finer models if necessary. Experiments show that this strategy results in significant saving in computational power with only negligible lose in context recognition accuracy.
  • Keywords
    audio signal processing; feature extraction; hidden Markov models; signal classification; adaptive context recognition; audio based context recognition; auditory feature extraction; classification algorithms; coarse HMM; coarse-to-fine training scheme; feature analysis; Cepstral analysis; Data mining; Feature extraction; Hidden Markov models; Mel frequency cepstral coefficient; Monitoring; Pattern recognition; Recurrent neural networks; Support vector machine classification; Support vector machines;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Pattern Recognition, 2008. ICPR 2008. 19th International Conference on
  • Conference_Location
    Tampa, FL
  • ISSN
    1051-4651
  • Print_ISBN
    978-1-4244-2174-9
  • Electronic_ISBN
    1051-4651
  • Type

    conf

  • DOI
    10.1109/ICPR.2008.4761905
  • Filename
    4761905