Title
Universal Adversarial Attacks On Text Classifiers
Abstract
Despite the vast success neural networks have achieved in different application domains, they have been proven to be vulnerable to adversarial perturbations (small changes in the input), which lead them to produce the wrong output. In this paper, we propose a novel method, based on gradient projection, for generating universal adversarial perturbations for text; namely sequence of words that can be added to any input in order to fool the classifier with high probability. We observed that text classifiers are quite vulnerable to such perturbations: inserting even a single adversarial word to the beginning of every input sequence can drop the accuracy from 93% to 50%.
Year
DOI
Venue
2019
10.1109/ICASSP.2019.8682430
2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP)
Keywords
Field
DocType
neural network, universal adversarial perturbation, gradient projection, text classifier
Pattern recognition,Computer science,Context model,Robustness (computer science),Gradient projection,Artificial intelligence,Probabilistic logic,Classifier (linguistics),Text categorization,Artificial neural network,Adversarial system
Conference
ISSN
Citations 
PageRank 
1520-6149
0
0.34
References 
Authors
0
4
Name
Order
Citations
PageRank
Melika Behjati100.34
Seyed-Mohsen Moosavi-Dezfooli262726.32
Mahdieh Soleymani Baghshah318817.78
Pascal Frossard4193.50