Title
ThiNet: Pruning CNN Filters for a Thinner Net.
Abstract
This paper aims at accelerating and compressing deep neural networks to deploy CNN models into small devices like mobile phones or embedded gadgets. We focus on filter level pruning, i.e., the whole filter will be discarded if it is less important. An effective and unified framework, ThiNet (stands for "Thin Net"), is proposed in this paper. We formally establish filter pruning as an optimization problem, and reveal that we need to prune filters based on statistics computed from its next layer, not the current layer, which differentiates ThiNet from existing methods. We also propose "gcos" (Group COnvolution with Shuffling), a more accurate group convolution scheme, to further reduce the pruned model size. Experimental results demonstrate the effectiveness of our method, which has advanced the state-of-the-art. Moreover, we show that the original VGG-16 model can be compressed into a very small model (ThiNet-Tiny) with only 2.66MB model size, but still preserve AlexNet level accuracy. This small model is evaluated on several benchmarks with different vision tasks (e.g., classification, detection, segmentation), and shows excellent generalization ability.
Year
DOI
Venue
2019
10.1109/TPAMI.2018.2858232
IEEE transactions on pattern analysis and machine intelligence
Keywords
Field
DocType
Convolution,Computational modeling,Task analysis,Acceleration,Training,Neural networks,Image coding
Task analysis,Pattern recognition,Computer science,Segmentation,Convolution,Shuffling,Artificial intelligence,Acceleration,Artificial neural network,Optimization problem,Pruning
Journal
Volume
Issue
ISSN
41
10
1939-3539
Citations 
PageRank 
References 
16
0.65
15
Authors
6
Name
Order
Citations
PageRank
Jian-Hao Luo1875.00
Hao Zhang220758.59
Hong-Yu Zhou3160.99
Chen-Wei Xie4893.07
Jianxin Wu53276154.17
Weiyao Lin673268.05