Title
Adversarial Robustness in Deep Learning: Attacks on Fragile Neurons
Abstract
We identify fragile and robust neurons of deep learning architectures using nodal dropouts of the first convolutional layer. Using an adversarial targeting algorithm, we correlate these neurons with the distribution of adversarial attacks on the network. Adversarial robustness of neural networks has gained significant attention in recent times and highlights an intrinsic weaknesses of deep learning networks against carefully constructed distortion applied to input images. In this paper, we evaluate the robustness of state-of-the-art image classification models trained on the MNIST and CIFAR10 datasets against the fast gradient sign method attack, a simple yet effective method of deceiving neural networks. Our method identifies the specific neurons of a network that are most affected by the adversarial attack being applied. We, therefore, propose to make fragile neurons more robust against these attacks by compressing features within robust neurons and amplifying the fragile neurons proportionally.
Year
DOI
Venue
2021
10.1007/978-3-030-86362-3_2
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT I
Keywords
DocType
Volume
Deep learning, Fragile neurons, Data perturbation, Adversarial targeting, Robustness analysis, Adversarial robustness
Conference
12891
ISSN
Citations 
PageRank 
0302-9743
0
0.34
References 
Authors
0
4
Name
Order
Citations
PageRank
Chandresh Pravin100.34
Martino Ivan201.01
G. Nicosia342.44
Varun Kumar Ojha400.34