DocumentCode
3422201
Title
A self-organizing neural model for context-based action recognition
Author
Kuniyoshi, Yasuo ; Shimozaki, Moriaki
Author_Institution
Sch. of Inf. Sci. & Technol., Univ. of Tokyo, Japan
fYear
2003
fDate
20-22 March 2003
Firstpage
442
Lastpage
445
Abstract
An artificial neural network model for visual recognition of actions is proposed. In contrast with the existing gesture recognition systems, our model learns to recognize "true" actions, i.e. object-directed actions with causal chains of events, such as "He threw the ball at the window and broke it". The core mechanism of our model consists of a triplet of parallel self-organizing networks; The first pair of networks learn and recognize "spatial relationship" and "movement patterns", whose output is integrated by the third "temporal context" network.
Keywords
cognitive systems; gesture recognition; self-organising feature maps; artificial neural network model; context-based action recognition; gesture recognition systems; movement patterns; object-directed actions; parallel self-organizing networks; self-organizing neural model; spatial relationship; visual recognition; Artificial neural networks; Context modeling; Educational robots; Hidden Markov models; Information science; Intelligent robots; Layout; Pattern recognition; Robot vision systems; Self-organizing networks;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Engineering, 2003. Conference Proceedings. First International IEEE EMBS Conference on
Print_ISBN
0-7803-7579-3
Type
conf
DOI
10.1109/CNE.2003.1196856
Filename
1196856
Link To Document