Title
Boosting The Transferability Of Adversarial Samples Via Attention
Abstract
The widespread deployment of deep models necessitates the assessment of model vulnerability in practice, especially for safety- and security-sensitive domains such as autonomous driving and medical diagnosis. Transfer-based attacks against image classifiers thus elicit mounting interest, where attackers are required to craft adversarial images based on local proxy models without the feedback information from remote target ones. However, under such a challenging but practical setup, the synthesized adversarial samples often achieve limited success due to overfitting to the local model employed. In this work, we propose a novel mechanism to alleviate the overfitting issue. It computes model attention over extracted features to regularize the search of adversarial examples, which prioritizes the corruption of critical features that are likely to be adopted by diverse architectures. Consequently, it can promote the transferability of resultant adversarial instances. Extensive experiments on ImageNet classifiers confirm the effectiveness of our strategy and its superiority to state-of-the-art benchmarks in both white-box and black-box settings.
Year
DOI
Venue
2020
10.1109/CVPR42600.2020.00124
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR)
DocType
ISSN
Citations 
Conference
1063-6919
0
PageRank 
References 
Authors
0.34
31
7
Name
Order
Citations
PageRank
Weibin Wu160.81
Yuxin Su2195.72
Xixian Chen32611.28
Shenglin Zhao41237.86
Irwin King56751325.94
Michael R. Lyu610985529.03
Yu-Wing Tai7202892.75