Title
Adversarial image detection in deep neural networks.
Abstract
Deep neural networks are more and more pervading many computer vision applications and in particular image classification. Notwithstanding that, recent works have demonstrated that it is quite easy to create adversarial examples, i.e., images malevolently modified to cause deep neural networks to fail. Such images contain changes unnoticeable to the human eye but sufficient to mislead the network. This represents a serious threat for machine learning methods. In this paper, we investigate the robustness of the representations learned by the fooled neural network, analyzing the activations of its hidden layers. Specifically, we tested scoring approaches used for kNN classification, in order to distinguish between correctly classified authentic images and adversarial examples. These scores are obtained searching only between the very same images used for training the network. The results show that hidden layers activations can be used to reveal incorrect classifications caused by adversarial attacks.
Year
DOI
Venue
2019
10.1007/s11042-018-5853-4
Multimedia Tools Appl.
Keywords
Field
DocType
Adversarial images detection, Deep convolutional neural network, Machine learning security
Pattern recognition,Image detection,Computer science,Robustness (computer science),Artificial intelligence,Contextual image classification,Artificial neural network,Deep neural networks,Adversarial system
Journal
Volume
Issue
ISSN
78
3
1573-7721
Citations 
PageRank 
References 
2
0.40
33
Authors
5
Name
Order
Citations
PageRank
Fabio Carrara1298.17
Fabrizio Falchi245955.65
Roberto Caldelli348137.01
Giuseppe Amato4573.87
Rudy Becarelli51065.97