Title
Learning Temporal Information From Spatial Information Using Capsnets For Human Action Recognition
Abstract
Capsule Networks (CapsNets) are recently introduced to overcome some of the shortcomings of traditional Convolutional Neural Networks (CNNs). CapsNets replace neurons in CNNs with vectors to retain spatial relationships among the features. In this paper, we propose a CapsNet architecture that employs individual video frames for human action recognition without explicitly extracting motion information. We also propose weight pooling to reduce the computational complexity and improve the classification accuracy by appropriately removing some of the extracted features. We show how the capsules of the proposed architecture can encode temporal information by using the spatial features extracted from several video frames. Compared with a traditional CNN of the same complexity, the proposed CapsNet improves action recognition performance by 12.11% and 22.29% on the KTH and UCF-sports datasets, respectively.
Year
DOI
Venue
2019
10.1109/icassp.2019.8683720
2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP)
Keywords
Field
DocType
CapsNet, Human Action Recognition, CNN, routing-by-agreement
Spatial analysis,ENCODE,Pattern recognition,Computer science,Convolutional neural network,Pooling,Action recognition,Artificial intelligence,Computational complexity theory
Conference
ISSN
Citations 
PageRank 
1520-6149
0
0.34
References 
Authors
0
3
Name
Order
Citations
PageRank
Abdullah M. Algamdi110.68
Victor Sanchez214431.22
Chang-Tsun Li3245.11