Title
Appending Adversarial Frames For Universal Video Attack
Abstract
This paper investigates the problem of generating adversarial examples for video classification. We project all videos onto a semantic space and a perception space, and point out that adversarial attack is to find a counterpart which is close to the target in the perception space but far from the target in the semantic space. Based on this formulation, we notice that conventional attacking methods mostly used Euclidean distance to measure the perception space, but we propose to make full use of the property of videos and assume a modified video with a few consecutive frames replaced by dummy contents (e.g., a black frame with texts of `thank you for watching' on it) to be close to the original video in the perception space though they have a large Euclidean gap. This leads to a new attack approach which only adds perturbations on the newly-added frames. We show its high success rates in attacking six state-of-the-art video classification networks, as well as its universality, i.e., transferring well across videos and models.
Year
DOI
Venue
2021
10.1109/WACV48630.2021.00324
2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WACV 2021
DocType
ISSN
Citations 
Conference
2472-6737
0
PageRank 
References 
Authors
0.34
0
5
Name
Order
Citations
PageRank
Chen Zhikai100.34
Ling-Xi Xie242937.79
Shanmin Pang34013.47
Yong He47812.64
Tian Qi500.34