Title
Differentiable Quantization of Deep Neural Networks.
Abstract
We propose differentiable quantization (DQ) for efficient deep neural network (DNN) inference where gradient descent is used to learn the quantizer's step size, dynamic range and bitwidth. Training with differentiable quantizers brings two main benefits: first, DQ does not introduce hyperparameters; second, we can learn for each layer a different step size, dynamic range and bitwidth. Our experiments show that DNNs with heterogeneous and learned bitwidth yield better performance than DNNs with a homogeneous one. Further, we show that there is one natural DQ parametrization especially well suited for training. We confirm our findings with experiments on CIFAR-10 and ImageNet and we obtain quantized DNNs with learned quantization parameters achieving state-of-the-art performance.
Year
Venue
DocType
2019
arXiv: Learning
Journal
Volume
Citations 
PageRank 
abs/1905.11452
1
0.35
References 
Authors
0
8
Name
Order
Citations
PageRank
Stefan Uhlich1357.62
Lukas Mauch2134.97
Kazuki Yoshiyama341.46
Fabien Cardinaux427919.00
Javier Alonso García541.46
Stephen Tiedemann641.46
Thomas Kemp724630.93
Akio Nakamura86214.45