Title
Block change learning for knowledge distillation
Abstract
Deep neural networks perform well but require high-performance hardware for their use in real-world environments. Knowledge distillation is a simple method for improving the performance of a small network by using the knowledge of a large complex network. Small and large networks are referred to as student and teacher models, respectively. Previous knowledge distillation approaches perform well in a relatively small teacher network (20–30 layers) but poorly in large teacher networks (50 layers). Here, we propose an approach called block change learning that performs local and global knowledge distillation by changing blocks comprised of layers. The method focuses on the knowledge transfer without losing information in a large teacher model, as the approach considers intra-relationships between layers using local knowledge distillation and inter-relationships between corresponding blocks. The results are demonstrated this approach as superior to state-of-the-art methods using feature extraction datasets (Market1501 and DukeMTMC-relD) and object classification datasets (CIFAR-100 and Caltech256). Furthermore, we showed that the performance of the proposed approach was superior to that of a fine-tuning approach using pretrained models.
Year
DOI
Venue
2020
10.1016/j.ins.2019.10.074
Information Sciences
Keywords
Field
DocType
Knowledge distillation,Model compression,Convolutional neural network
Large networks,Knowledge transfer,Feature extraction,Distillation,Artificial intelligence,Complex network,Mathematics,Deep neural networks,Machine learning
Journal
Volume
ISSN
Citations 
513
0020-0255
0
PageRank 
References 
Authors
0.34
0
4
Name
Order
Citations
PageRank
Hyunguk Choi111.03
Younkwan Lee222.73
Kin Choong Yow326462.38
Moongu Jeon445672.81