Title
Accelerating Neural Architecture Search for Natural Language Processing with Knowledge Distillation and Earth Mover's Distance
Abstract
ABSTRACTRecent AI research has witnessed increasing interests in automatically designing the architecture of deep neural networks, which is coined as neural architecture search (NAS). The automatically searched network architectures via NAS methods have outperformed manually designed architectures on some NLP tasks. However, training a large number of model configurations for efficient NAS is computationally expensive, creating a substantial barrier for applying NAS methods in real-life applications. In this paper, we propose to accelerate neural architecture search for natural language processing based on knowledge distillation (called KD-NAS). Specifically, instead of searching the optimal network architecture on the validation set conditioned on the optimal network weights on the training set, we learn the optimal network by minimizing the knowledge loss transferred from a pre-trained teacher network to the searching network based on Earth Mover's Distance (EMD). Experiments on five datasets show that our method achieves promising performance compared to strong competitors on both accuracy and searching speed. For reproducibility, we submit the code at: https://github.com/lxk00/KD-NAS-EMD.
Year
DOI
Venue
2021
10.1145/3404835.3463017
Research and Development in Information Retrieval
Keywords
DocType
Citations 
Neural architecture search, knowledge distillation, earth mover's distance
Conference
0
PageRank 
References 
Authors
0.34
0
6
Name
Order
Citations
PageRank
Jianquan Li175.26
Xiaokang Liu201.01
Sheng Zhang361.44
Min Yang47720.41
Xu Ruifeng543253.04
Feng-qing Qin6221.88