Title
Compression Of Acoustic Model Via Knowledge Distillation And Pruning
Abstract
Recently, the performance of speech recognition system based on neural network has been greatly improved. Arguably, this huge improvement can be mainly attributed to deeper and wider layers. These systems are more difficult to be deployed on the embedded devices due to their large size and high computational complexity. To address these issues, we propose a method to compress deep feed-forward neural network (DNN) based acoustic model. In detail, a state-of-the-art acoustic model is trained as the baseline model. In this step, layer normalization is applied to accelerating the model convergence and improving the generalization performance. Knowledge distillation and pruning are then conducted to compress the model. Our final model can achieve 14:59x parameters reduction, 5x storage size reduction and comparable performance compared with the baseline model.
Year
DOI
Venue
2018
10.1109/ICPR.2018.8545028
2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR)
Keywords
Field
DocType
speech recognition, layer normalization, knowledge distillation, pruning, model compression
Convergence (routing),Normalization (statistics),Pattern recognition,Computer science,Recurrent neural network,Distillation,Artificial intelligence,Artificial neural network,Computer engineering,Acoustic model,Computational complexity theory,Fold (higher-order function)
Conference
ISSN
Citations 
PageRank 
1051-4651
0
0.34
References 
Authors
0
5
Name
Order
Citations
PageRank
Chenxing Li1146.76
Lei Zhu2734.38
Shuang Xu327432.53
Peng Gao400.68
Bo Xu51309.43