DocumentCode :
249531
Title :
Learning a sparse dictionary of video structure for activity modeling
Author :
Nayak, N.M. ; Roy-Chowdhury, A.K.
Author_Institution :
Univ. of California, Riverside, Riverside, CA, USA
fYear :
2014
fDate :
27-30 Oct. 2014
Firstpage :
4892
Lastpage :
4896
Abstract :
We present an approach which incorporates spatiotemporal features as well as the relationships between them, into a sparse dictionary learning framework for activity recognition. We propose that the dictionary learning framework can be adapted to learning complex relationships between features in an unsupervised manner. From a set of training videos, a dictionary is learned for individual features, as well as the relationships between them using a stacked predictive sparse decomposition framework. This combined dictionary provides a representation of the structure of the video and is spatio-temporally pooled in a local manner to obtain descriptors. The descriptors are then combined using a multiple kernel learning framework to design classifiers. Experiments have been conducted on two popular activity recognition datasets to demonstrate the superior performance of our approach on single person as well as multi-person activities.
Keywords :
feature extraction; image classification; signal processing; spatiotemporal phenomena; unsupervised learning; video signal processing; activity modeling; activity recognition datasets; multiple kernel learning framework; sparse dictionary learning framework; spatiotemporal features; stacked predictive sparse decomposition framework; video structure; video training; Computer vision; Dictionaries; Encoding; Kernel; Pattern recognition; Spatiotemporal phenomena; Training; Sparse coding; activity recognition; multiple kernel learning;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Image Processing (ICIP), 2014 IEEE International Conference on
Conference_Location :
Paris
Type :
conf
DOI :
10.1109/ICIP.2014.7025991
Filename :
7025991
Link To Document :
بازگشت