Title
Loss-aware Binarization of Deep Networks.
Abstract
Deep neural network models, though very powerful and highly successful, are computationally expensive in terms of space and time. Recently, there have been a number of attempts on binarizing the network weights and activations. This greatly reduces the network size, and replaces the underlying multiplications to additions or even XNOR bit operations. However, existing binarization schemes are based on simple matrix approximations and ignore the effect of binarization on the loss. In this paper, we propose a proximal Newton algorithm with diagonal Hessian approximation that directly minimizes the loss w.r.t. the binarized weights. The underlying proximal step has an efficient closed-form solution, and the second-order information can be efficiently obtained from the second moments already computed by the Adam optimizer. Experiments on both feedforward and recurrent networks show that the proposed loss-aware binarization algorithm outperforms existing binarization schemes, and is also more robust for wide and deep networks.
Year
Venue
Field
2016
international conference on learning representations
Diagonal,Network size,Mathematical optimization,XNOR gate,Computer science,Matrix (mathematics),Spacetime,Hessian matrix,Artificial intelligence,Artificial neural network,Machine learning,Feed forward
DocType
Volume
Citations 
Journal
abs/1611.01600
7
PageRank 
References 
Authors
0.42
0
3
Name
Order
Citations
PageRank
lu hou1626.80
Quanming Yao228827.13
James T. Kwok34920312.83