Abstract | ||
---|---|---|
The motion of a video contains two factors: magnitude and direction, but most of the existing video self-supervised methods ignored the motion direction information. In this paper, we propose a Video Motion Perception (VMP) self-supervised framework, simultaneously taking account of the above two key factors. Specifically, a Motion Direction Perception Module (MDPM) is applied to asking the network to predict the moving direction of the video objects by using two well-designed handcraft strategies. Additionally, we analyze the characteristic of video motion in natural scenes and propose the Motion Change Perception Module (MCPM) accordingly for motion magnitude learning. Experimental results show that VMP achieves competitive performance on different benchmarks, including action recognition, video retrieval, and action similarity labeling. |
Year | DOI | Venue |
---|---|---|
2022 | 10.1007/978-3-031-15937-4_43 | ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2022, PT IV |
Keywords | DocType | Volume |
Self-supervised learning, Action recognition, Video motion perception | Conference | 13532 |
ISSN | Citations | PageRank |
0302-9743 | 0 | 0.34 |
References | Authors | |
0 | 6 |
Name | Order | Citations | PageRank |
---|---|---|---|
Wei Li | 1 | 0 | 0.34 |
Dezhao Luo | 2 | 0 | 0.34 |
Bo Fang | 3 | 0 | 0.34 |
Xiaoni Li | 4 | 0 | 1.35 |
Yu Zhou | 5 | 13 | 2.18 |
Weiping Wang | 6 | 7 | 9.20 |