DocumentCode :
1832458
Title :
Spatial-temporal motion information integration for action detection and recognition in non-static background
Author :
Dianting Liu ; Mei-Ling Shyu ; Guiru Zhao
Author_Institution :
Dept. of Electr. & Comput. Eng., Univ. of Miami, Coral Gables, FL, USA
fYear :
2013
fDate :
14-16 Aug. 2013
Firstpage :
626
Lastpage :
633
Abstract :
Various motion detection methods have been proposed in the past decade, but there are seldom attempts to investigate the advantages and disadvantages of different detection mechanisms so that they can complement each other to achieve a better performance. Toward such a demand, this paper proposes a human action detection and recognition framework to bridge the semantic gap between low-level pixel intensity change and the high-level understanding of the meaning of an action. To achieve a robust estimation of the region of action with the complexities of an uncontrolled background, we propose the combination of the optical flow field and Harris3D corner detector to obtain a new spatial-temporal estimation in the video sequences. The action detection method, considering the integrated motion information, works well with the dynamic background and camera motion, and demonstrates the advantage of the proposed method of integrating multiple spatial-temporal cues. Then the local features (SIFT and STIP) extracted from the estimated region of action are used to learn the Universal Background Model (UBM) for the action recognition task. The experimental results on KTH and UCF YouTube Action (UCF11) data sets show that the proposed action detection and recognition framework can not only better estimate the region of action but also achieve better recognition accuracy comparing with the peer work.
Keywords :
feature extraction; gesture recognition; image sequences; motion estimation; transforms; Harris3D corner detector; KTH data set; SIFT; STIP; UBM; UCF YouTube Action data set; UCF11 data set; action meaning; camera motion; dynamic background; high-level understanding; human action detection; human action recognition; local feature extraction; low-level pixel intensity change; motion detection method; multiple spatial-temporal cue integration; nonstatic background; optical flow field; robust region of action estimation; semantic gap; spatial-temporal motion information integration; universal background model; video sequences; Cameras; Detectors; Estimation; Feature extraction; Optical imaging; Vectors; Video sequences; Action Detection; Action Recognition; GMM Supervector; Gaussian Mixture Models (GMM); Spatio-temporal Motion Information Integration; Universal Background Model (UBM);
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Information Reuse and Integration (IRI), 2013 IEEE 14th International Conference on
Conference_Location :
San Francisco, CA
Type :
conf
DOI :
10.1109/IRI.2013.6642527
Filename :
6642527
Link To Document :
بازگشت