Abstract | ||
---|---|---|
Deep neural networks have achieved outstanding progress in many fields, such as computer vision, speech recognition and natural language processing. However, large deep neural networks often need huge storage space and long training time, making them difficult to apply to resource restricted devices. In this paper, we propose a method for compressing the structure of deep neural networks. Specifically, we apply clustering analysis to find similar neurons in each layer of the original network, and merge them and the corresponding connections. After the compression of the network, the number of parameters in the deep neural network is significantly reduced, and the required storage space and computational time is greatly reduced as well. We test our method on deep belief network (DBN) and two convolutional neural networks. The experimental results demonstrate that our proposed method can greatly reduce the number of parameters of the deep networks, while keeping their classification accuracy. Especially, on the CIFAR-10 dataset, we have compressed VGGNet with compression ratio 92.96%, and the final model after fine-tuning obtains even higher accuracy than the original model. |
Year | DOI | Venue |
---|---|---|
2020 | 10.1007/s12559-019-09703-6 | Cognitive Computation |
Keywords | DocType | Volume |
Machine learning, Deep neural networks, Structure compression, Neurons, Clustering | Journal | 12 |
Issue | ISSN | Citations |
3 | 1866-9956 | 0 |
PageRank | References | Authors |
0.34 | 0 | 6 |
Name | Order | Citations | PageRank |
---|---|---|---|
Guoqiang Zhong | 1 | 123 | 20.68 |
Wenxue Liu | 2 | 0 | 1.35 |
Hui Yao | 3 | 0 | 0.68 |
Tao Li | 4 | 0 | 0.34 |
Jinxuan Sun | 5 | 3 | 1.40 |
Xiang Liu | 6 | 0 | 1.35 |