Title | ||
---|---|---|
Eager pruning: algorithm and architecture support for fast training of deep neural networks |
Abstract | ||
---|---|---|
Today's big and fast data and the changing circumstance require fast training of Deep Neural Networks (DNN) in various applications. However, training a DNN with tons of parameters involves intensive computation. Enlightened by the fact that redundancy exists in DNNs and the observation that the ranking of the significance of the weights changes slightly during training, we propose Eager Pruning, which speeds up DNN training by moving pruning to an early stage.
Eager Pruning is supported by an algorithm and architecture co-design. The proposed algorithm dictates the architecture to identify and prune insignificant weights during training without accuracy loss. A novel architecture is designed to transform the reduced training computation into performance improvement. Our proposed Eager Pruning system gains an average of 1.91x speedup over state-of-the-art hardware accelerator and 6.31x energy-efficiency over Nvidia GPUs.
|
Year | DOI | Venue |
---|---|---|
2019 | 10.1145/3307650.3322263 | Proceedings of the 46th International Symposium on Computer Architecture |
Keywords | Field | DocType |
neural network pruning, neural network training, software-hardware co-design | Architecture,Ranking,Computer science,Parallel computing,Redundancy (engineering),Hardware acceleration,Artificial intelligence,Machine learning,Speedup,Pruning,Computation,Performance improvement | Conference |
ISSN | ISBN | Citations |
1063-6897 | 978-1-4503-6669-4 | 7 |
PageRank | References | Authors |
0.55 | 18 | 4 |
Name | Order | Citations | PageRank |
---|---|---|---|
Zhang, Jiaqi | 1 | 73 | 11.73 |
Xiangru Chen | 2 | 7 | 0.88 |
Mingcong Song | 3 | 24 | 1.20 |
Tao Li | 4 | 761 | 47.52 |