Title
Adapting Neural Architectures Between Domains
Abstract
Neural architecture search (NAS) has demonstrated impressive performance in automatically designing high-performance neural networks. The power of deep neural networks is to be unleashed for analyzing a large volume of data (e.g. ImageNet), but the architecture search is often executed on another smaller dataset (e.g. CIFAR-10) to finish it in a feasible time. However, it is hard to guarantee that the optimal architecture derived on the proxy task could maintain its advantages on another more challenging dataset. This paper aims to improve the generalization of neural architectures via domain adaptation. We analyze the generalization bounds of the derived architecture and suggest its close relations with the validation error and the data distribution distance on both domains. These theoretical analyses lead to AdaptNAS, a novel and principled approach to adapt neural architectures between domains in NAS. Our experimental evaluation shows that only a small part of ImageNet will be sufficient for AdaptNAS to extend its architecture success to the entire ImageNet and outperform state-of-the-art comparison algorithms.
Year
Venue
DocType
2020
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS (NEURIPS 2020)
Conference
Volume
ISSN
Citations 
33
1049-5258
0
PageRank 
References 
Authors
0.34
0
4
Name
Order
Citations
PageRank
Yanxi Li102.70
Zhaohui Yang2355.01
Yunhe Wang311322.76
Chang Xu478147.60