• DocumentCode
    607770
  • Title

    Recognizing human actions from noisy videos via multiple instance learning

  • Author

    Sener, Fadime ; Samet, N. ; Duygulu, P. ; Ikizler-Cinbis, N.

  • Author_Institution
    Bilgisayar Muhendisligi Bolumu, Bilkent Univ., Ankara, Turkey
  • fYear
    2013
  • fDate
    24-26 April 2013
  • Firstpage
    1
  • Lastpage
    4
  • Abstract
    In this work, we study the task of recognizing human actions from noisy videos and effects of noise to recognition performance and propose a possible solution. Datasets available in computer vision literature are relatively small and could include noise due to labeling source. For new and relatively big datasets, noise amount would possible increase and the performance of traditional instance based learning methods is likely to decrease. In this work, we propose a multiple instance learning-based solution in case of an increase in noise. For this purpose, each video is represented with spatio-temporal features, then bag-of-words method is applied. Then, using support vector machines (SVM), both instance-based learning and multiple instance learning classifiers are constructed and compared. The classification results show that multiple instance learning classifiers has better performance than instance based learning counterparts on noisy videos.
  • Keywords
    computer vision; image representation; learning (artificial intelligence); support vector machines; video signal processing; SVM; bag-of-words method; computer vision; human action recognition; instance learning classifier; noisy video; spatio-temporal feature; support vector machine; video representation; Bismuth; Computer vision; Hidden Markov models; Histograms; Noise; Noise measurement; Videos; Data Noise; Human Action Recognition; Multiple Instance Learning; Video Understanding;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Signal Processing and Communications Applications Conference (SIU), 2013 21st
  • Conference_Location
    Haspolat
  • Print_ISBN
    978-1-4673-5562-9
  • Electronic_ISBN
    978-1-4673-5561-2
  • Type

    conf

  • DOI
    10.1109/SIU.2013.6531431
  • Filename
    6531431