Title
CHaPR: Efficient Inference of CNNs via Channel Pruning
Abstract
To deploy a CNN on resource-constrained edge platforms, channel pruning techniques promise a significant reduction of implementation costs including memory, computation, and energy consumption without special hardware or software libraries. This paper proposes CHaPR, a novel pruning technique to structurally prune the redundant channels in a trained deep Convolutional Neural Network. CHaPR utilizes a proposed subset selection problem formulation for pruning which it solves using pivoted QR factorization. CHaPR also includes an additional pruning technique for ResNet-like architectures which resolves the issue encountered by some existing channel pruning methods that not all the layers can be pruned. Experimental results on VGG-16 and ResNet-50 models show 4.29X and 2.84X reduction, respectively in computation cost while incurring 2.50% top-1 and 1.40% top-5 accuracy losses. Compared to many existing works, CHaPR performs better when considering an Overall Score metric which accounts for both computation and accuracy.
Year
DOI
Venue
2020
10.1109/COINS49042.2020.9191636
2020 International Conference on Omni-layer Intelligent Systems (COINS)
Keywords
DocType
ISBN
Convolutional Neural Networks,Model Pruning
Conference
978-1-7281-6372-7
Citations 
PageRank 
References 
0
0.34
2
Authors
3
Name
Order
Citations
PageRank
boyu zhang17117.54
Azadeh Davoodi236234.99
Yu Hen Hu31398157.52