Title
GCF-Net: Gated Clip Fusion Network for Video Action Recognition
Abstract
In recent years, most of the accuracy gains for video action recognition have come from the newly designed CNN architectures (e.g., 3D-CNNs). These models are trained by applying a deep CNN on single clip of fixed temporal length. Since each video segment are processed by the 3D-CNN module separately, the corresponding clip descriptor is local and the inter-clip relationships are inherently implicit. Common method that directly averages the clip-level outputs as a video-level prediction is prone to fail due to the lack of mechanism that can extract and integrate relevant information to represent the video. In this paper, we introduce the Gated Clip Fusion Network (GCF-Net) that can greatly boost the existing video action classifiers with the cost of a tiny computation overhead. The GCF-Net explicitly models the inter-dependencies between video clips to strengthen the receptive field of local clip descriptors. Furthermore, the importance of each clip to an action event is calculated and a relevant subset of clips is selected accordingly for a video-level analysis. On a large benchmark dataset (Kinetics-600), the proposed GCF-Net elevates the accuracy of existing action classifiers by 11.49% (based on central clip) and 3.67% (based on densely sampled clips) respectively.
Year
DOI
Venue
2020
10.1007/978-3-030-68238-5_46
ECCV Workshops
Keywords
DocType
Citations 
Video action recognition,3D-CNNs,Dense slip sampling,Clip fusion
Conference
2
PageRank 
References 
Authors
0.41
1
3
Name
Order
Citations
PageRank
Jenhao Hsiao121.42
Jiawei Chen28714.73
Chiuman Ho321.42