Title
Fixed-Point Back-Propagation Training
Abstract
Recent emerged quantization technique (i.e., using low bit-width fixed-point data instead of high bit-width floating-point data) has been applied to inference of deep neural networks for fast and efficient execution. However, directly applying quantization in training can cause significant accuracy loss, thus remaining an open challenge. In this paper; we propose a novel training approach, which applies a layer-wise precision-adaptive quantization in deep neural networks. The new training approach leverages our key insight that the degradation of training accuracy is attributed to the dramatic change of data distribution. Therefore, by keeping the data distribution stable through a layer-wise precision-adaptive quantization, we are able to directly train deep neural networks using low bit-width fixed-point data and achieve guaranteed accuracy, without changing hyper parameters. Experimental results on a wide variety of network architectures (e.g., convolution and recurrent networks) and applications (e.g., image classification, object detection, segmentation and machine translation) show that the proposed approach can train these neural networks with negligible accuracy losses (-1.40%similar to 1.3%, 0.02% on average), and speed up training by 252% on a state-of-the-art Intel CPU.
Year
DOI
Venue
2020
10.1109/CVPR42600.2020.00240
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR)
DocType
ISSN
Citations 
Conference
1063-6919
0
PageRank 
References 
Authors
0.34
0
11
Name
Order
Citations
PageRank
Xishan Zhang172.90
Shaoli Liu256027.88
Rui Zhang301.01
Chang Liu4571117.41
Di Huang522.19
Shiyi Zhou600.34
Jiaming Guo712.38
Qi Guo871634.09
Zidong Du957429.68
Tian Zhi1000.68
Yunji Chen11143279.99