DocumentCode
3188459
Title
Action Recognition by Local Space-Time Features and Least Square Twin SVM (LS-TSVM)
Author
Mozafari, Kourosh ; Nasir, Jalal A. ; Charkar, Nasrollah Moghadam ; Jalili, Saeed
Author_Institution
Dept. of Electr. & Comput. Eng., Tarbiat Modares Univ., Tehran, Iran
fYear
2011
fDate
12-14 Dec. 2011
Firstpage
287
Lastpage
292
Abstract
In this research a new approach for human action recognition is proposed. At first, local space-time features extracted which recently becomes a popular video representation. Feature extraction is done with use of Harris detector algorithm and Histogram of Optical Flow (HOF) descriptor. Then we apply a new extended SVM classifier called least square Twin SVM (LS-TSVM). LS-TSVM is a binary classifier that does classification by use of two nonparallel hyperplanes and it is four times faster than the classical SVM while the precision is better. We investigate the performance of LS-TSVM method on a total of 25 persons on KTH dataset. Our experiments on the standard KTH action dataset shown that our method improves state-of-the-art results by achieving 95.8%, 96.3% and 97.2%% accuracy in case of 1-fold , 5-fold and 10-fold cross validation.
Keywords
data handling; feature extraction; gesture recognition; image classification; image sequences; support vector machines; Harris detector algorithm; KTH action dataset; SVM classifier; histogram of optical flow descriptor; human action recognition; least square twin SVM; local space-time feature extraction; nonparallel hyperplane; video representation; Accuracy; Computer vision; Feature extraction; Histograms; Humans; Support vector machines; Training data; Action Recognition; Harris; Histogram of Optical Floow (HOF); KTH dataset.; LS-TSVM; Twin Support Vector Machines;
fLanguage
English
Publisher
ieee
Conference_Titel
Informatics and Computational Intelligence (ICI), 2011 First International Conference on
Conference_Location
Bandung
Print_ISBN
978-1-4673-0091-9
Type
conf
DOI
10.1109/ICI.2011.55
Filename
6141687
Link To Document