Title
Eager pruning: algorithm and architecture support for fast training of deep neural networks
Abstract
Today's big and fast data and the changing circumstance require fast training of Deep Neural Networks (DNN) in various applications. However, training a DNN with tons of parameters involves intensive computation. Enlightened by the fact that redundancy exists in DNNs and the observation that the ranking of the significance of the weights changes slightly during training, we propose Eager Pruning, which speeds up DNN training by moving pruning to an early stage. Eager Pruning is supported by an algorithm and architecture co-design. The proposed algorithm dictates the architecture to identify and prune insignificant weights during training without accuracy loss. A novel architecture is designed to transform the reduced training computation into performance improvement. Our proposed Eager Pruning system gains an average of 1.91x speedup over state-of-the-art hardware accelerator and 6.31x energy-efficiency over Nvidia GPUs.
Year
DOI
Venue
2019
10.1145/3307650.3322263
Proceedings of the 46th International Symposium on Computer Architecture
Keywords
Field
DocType
neural network pruning, neural network training, software-hardware co-design
Architecture,Ranking,Computer science,Parallel computing,Redundancy (engineering),Hardware acceleration,Artificial intelligence,Machine learning,Speedup,Pruning,Computation,Performance improvement
Conference
ISSN
ISBN
Citations 
1063-6897
978-1-4503-6669-4
7
PageRank 
References 
Authors
0.55
18
4
Name
Order
Citations
PageRank
Zhang, Jiaqi17311.73
Xiangru Chen270.88
Mingcong Song3241.20
Tao Li476147.52