Title
Exploiting Linear Structure Within Convolutional Networks for Efficient Evaluation.
Abstract
We present techniques for speeding up the test-time evaluation of large convolutional networks, designed for object recognition tasks. These models deliver impressive accuracy, but each image evaluation requires millions of floating point operations, making their deployment on smartphones and Internet-scale clusters problematic. The computation is dominated by the convolution operations in the lower layers of the model. We exploit the redundancy present within the convolutional filters to derive approximations that significantly reduce the required computation. Using large state-of-the-art models, we demonstrate speedups of convolutional layers on both CPU and GPU by a factor of 2x, while keeping the accuracy within 1% of the original model.
Year
Venue
DocType
2014
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 27 (NIPS 2014)
Conference
Volume
ISSN
Citations 
27
1049-5258
230
PageRank 
References 
Authors
16.58
9
5
Search Limit
100230
Name
Order
Citations
PageRank
Emily Denton137825.96
Wojciech Zaremba22733117.55
J. Bruna3169782.95
Yann LeCun4260903771.21
Robert Fergus511214735.18