Title
Temporal Shift Module for Efficient Video Understanding.
Abstract
The explosive growth in online video streaming gives rise to challenges on efficiently extracting the spatial-temporal information to perform video understanding. Conventional 2D CNNs are computationally cheap but cannot capture long-term temporal relationships; 3D CNN based methods can achieve good performance but are computationally intensive, making it expensive to deploy. In this paper, we propose a generic and effective Temporal Shift Module (TSM) that enjoys both high efficiency and high performance. Specifically, it can achieve the performance of 3D CNN but maintain 2D complexity. The central idea of TSM is to shift part of the channels along the temporal dimension, which facilitates information exchange among neighboring frames. TSM can be inserted into 2D CNNs to achieve temporal modeling at the cost of zero FLOPs and zero parameters. On the Something-Something-V1 dataset which focuses on temporal modeling, we achieved better results than I3D family and ECO family using 6X and 2.7X fewer FLOPs respectively. Measured on P100 GPU, our single model achieved 1.8% higher accuracy at 8X lower latency and 12X higher throughput compared to I3D. Remarkably, our framework ranks the first on both Something-Something V1 and V2 leaderboards upon this paperu0027s submission.
Year
Venue
Field
2018
arXiv: Computer Vision and Pattern Recognition
Object detection,Pattern recognition,Computer science,Explosive material,Video streaming,Communication channel,Real-time computing,Temporal modeling,Artificial intelligence,Online video,Latency (engineering),Computation
DocType
Volume
Citations 
Journal
abs/1811.08383
0
PageRank 
References 
Authors
0.34
22
3
Name
Order
Citations
PageRank
Lin, Ji1798.18
Chuang Gan204.39
Song Han3210279.81