Title
Sparse, Quantized, Full Frame CNN for Low Power Embedded Devices
Abstract
This paper presents methods to reduce the complexity of convolutional neural networks (CNN). These include: (1) A method to quickly and easily sparsify a given network. (2) Fine tune the sparse network to obtain the lost accuracy back (3) Quantize the network to be able to implement it using 8-bit fixed point multiplications efficiently. (4) We then show how an inference engine can be designed to take advantage of the sparsity. These techniques were applied to full frame semantic segmentation and the degradation due to the sparsity and quantization is found to be negligible. We show by analysis that the complexity reduction achieved is significant. Results of implementation on Texas Instruments TDA2x SoC [17] are presented. We have modified Caffe CNN framework to do the sparse, quantized training described in this paper. The source code for the training is made available at https://github.com/tidsp/caffe-jacinto.
Year
DOI
Venue
2017
10.1109/CVPRW.2017.46
2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)
Keywords
Field
DocType
full frame CNN,low power embedded devices,convolutional neural networks,sparse network,8-bit fixed point multiplications,full frame semantic segmentation,complexity reduction,Texas Instruments TDA2x SoC,Caffe CNN framework,source code
Computer vision,Digital signal processing,Computer science,Convolutional neural network,Source code,Caffè,Reduction (complexity),Inference engine,Artificial intelligence,Fixed point,Quantization (signal processing)
Conference
Volume
Issue
ISSN
2017
1
2160-7508
ISBN
Citations 
PageRank 
978-1-5386-0734-3
4
0.46
References 
Authors
12
4
Name
Order
Citations
PageRank
Manu Mathew153.18
Kumar Desappan281.88
Pramod Swami382.56
Soyeb Nagori4102.67