Title
Compressing Models with Few Samples: Mimicking then Replacing
Abstract
Few-sample compression aims to compress a big redundant model into a small compact one with only few samples. If we fine-tune models with these limited few samples directly, models will be vulnerable to overfit and learn almost nothing. Hence, previous methods optimize the compressed model layer-by-layer and try to make every layer have the same outputs as the corresponding layer in the teacher model, which is cumbersome. In this paper, we propose a new framework named Mimicking then Replacing (MiR) for few-sample compression, which firstly urges the pruned model to output the same features as the teacher's in the penultimate layer, and then replaces teacher's layers before penultimate with a well-tuned compact one. Unlike previous layer-wise reconstruction methods, our MiR optimizes the entire network holistically, which is not only simple and effective, but also unsupervised and general. MiR outperforms previous methods with large margins. Codes is available at https://github.com/cjnjuwhy/MiR.
Year
DOI
Venue
2022
10.1109/CVPR52688.2022.00078
2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
Keywords
DocType
Volume
Deep learning architectures and techniques, Representation learning, Transfer/low-shot/long-tail learning
Conference
2022
Issue
ISSN
ISBN
1
1063-6919
978-1-6654-6947-0
Citations 
PageRank 
References 
0
0.34
7
Authors
6
Name
Order
Citations
PageRank
Huanyu Wang100.34
Junjie Liu200.34
Xin Ma300.34
Yang Yong400.34
Zhenhua Chai5126.59
Jianxin Wu63276154.17