DocumentCode :
3748807
Title :
Action Localization in Videos through Context Walk
Author :
Khurram Soomro;Haroon Idrees;Mubarak Shah
Author_Institution :
Center for Res. in Comput. Vision, Univ. of Central Florida, Orlando, FL, USA
fYear :
2015
Firstpage :
3280
Lastpage :
3288
Abstract :
This paper presents an efficient approach for localizing actions by learning contextual relations, in the form of relative locations between different video regions. We begin by over-segmenting the videos into supervoxels, which have the ability to preserve action boundaries and also reduce the complexity of the problem. Context relations are learned during training which capture displacements from all the supervoxels in a video to those belonging to foreground actions. Then, given a testing video, we select a supervoxel randomly and use the context information acquired during training to estimate the probability of each supervoxel belonging to the foreground action. The walk proceeds to a new supervoxel and the process is repeated for a few steps. This "context walk" generates a conditional distribution of an action over all the supervoxels. A Conditional Random Field is then used to find action proposals in the video, whose confidences are obtained using SVMs. We validated the proposed approach on several datasets and show that context in the form of relative displacements between supervoxels can be extremely useful for action localization. This also results in significantly fewer evaluations of the classifier, in sharp contrast to the alternate sliding window approaches.
Keywords :
"Videos","Context","Training","Testing","Proposals","Electron tubes","Support vector machines"
Publisher :
ieee
Conference_Titel :
Computer Vision (ICCV), 2015 IEEE International Conference on
Electronic_ISBN :
2380-7504
Type :
conf
DOI :
10.1109/ICCV.2015.375
Filename :
7410732
Link To Document :
بازگشت