DocumentCode
248207
Title
Action classification by exploring directional co-occurrence of weighted stips
Author
Mengyuan Liu ; Hong Liu ; Qianru Sun
Author_Institution
Eng. Lab. on Intell. Perception for Internet of Things, Peking Univ., Shenzhen, China
fYear
2014
fDate
27-30 Oct. 2014
Firstpage
1460
Lastpage
1464
Abstract
Human action recognition is challenging mainly due to intro-variety, inter-ambiguity and clutter backgrounds in real videos. Bag-of-visual words model utilizes spatio-temporal interest points(STIPs), and represents action by the distribution of points which ignores visual context among points. To add more contextual information, we propose a method by encoding spatio-temporal distribution of weighted pairwise points. First, STIPs are extracted from an action sequence and clustered into visual words. Then, each word is weighted in both temporal and spatial domains to capture the relationships with other words. Finally, the directional relationships between co-occurrence pairwise words are used to encode visual contexts. We report state-of-the-art results on Rochester and UT-Interaction datasets to validate that our method can classify human actions with high accuracies.
Keywords
image classification; image coding; image sequences; Rochester datasets; STIP clustering; STIP extraction; UT-Interaction datasets; action classification; action sequence; bag-of-visual words; co-occurrence pairwise words; contextual information; directional relationships; human action recognition; spatiotemporal distribution; spatiotemporal interest points; visual context encoding; weighted pairwise points; Accuracy; Context; Feature extraction; Histograms; Sun; Videos; Visualization; Spatio-temporal interest point; bag-of-visual words; co-occurrence;
fLanguage
English
Publisher
ieee
Conference_Titel
Image Processing (ICIP), 2014 IEEE International Conference on
Conference_Location
Paris
Type
conf
DOI
10.1109/ICIP.2014.7025292
Filename
7025292
Link To Document