Title :
Automatic collection of Web video shots corresponding to specific actions using Web images
Author :
Nga, Do Hang ; Yanai, Keiji
Author_Institution :
Dept. of Inf., Univ. of Electro-Commun., Chofu, Japan
Abstract :
In this paper, we apply Web images to the problem of automatically extracting video shots corresponding to specific actions from Web videos. Our framework modifies the unsupervised method on automatic collecting of Web video shots corresponding to the given actions which we proposed last year [9]. For each action, following that work, we first exploit tag relevance to gather 200 most relevant videos of the given action and segment each video into several video shots. Shots are then converted into bags of spatio-temporal features and ranked by the VisualRank method. We refine the approach by introducing the use of Web action images into shot ranking step. We select images by applying Pose-lets [2] to detect human in the case of human actions. We test our framework on 28 human action categories whose precision values were 20% or below and 8 non-human action categories whose precision values were less than 15% in [9]. The results show that our model can improve the precision approximately 6% over 28 human action categories and 16% over 8 non-human action categories.
Keywords :
Internet; feature extraction; image segmentation; object detection; video retrieval; Pose-lets; VisualRank method; Web action images; Web video shots automatic collection; human detection; nonhuman action categories; shot ranking step; spatio-temporal features; specific actions; tag relevance exploitation; unsupervised method; video segmentation; video shot extraction; Feature extraction; Humans; Noise measurement; Training; Vectors; Visualization; YouTube;
Conference_Titel :
Computer Vision and Pattern Recognition Workshops (CVPRW), 2012 IEEE Computer Society Conference on
Conference_Location :
Providence, RI
Print_ISBN :
978-1-4673-1611-8
Electronic_ISBN :
2160-7508
DOI :
10.1109/CVPRW.2012.6239255