Title
Attention-Driven Loss for Anomaly Detection in Video Surveillance
Abstract
AbstractRecent video anomaly detection methods focus on reconstructing or predicting frames. Under this umbrella, the long-standing inter-class data-imbalance problem resorts to the imbalance between foreground and stationary background objects in video anomaly detection and this has been less investigated by existing solutions. Naively optimizing the reconstructing loss yields a biased optimization towards background reconstruction rather than the objects of interest in the foreground. To solve this, we proposed a simple yet effective solution, termed attention-driven loss to alleviate the foreground-background imbalance problem in anomaly detection. Specifically, we compute a single mask map that summarizes the frame evolution of moving foreground regions and suppresses the background in the training video clips. After that, we construct an attention map through the combination of the mask map and background to give different weights to the foreground and background region respectively. The proposed attention-driven loss is independent of backbone networks and can be easily augmented in most existing anomaly detection models. Augmented with attention-driven loss, the model is able to achieve AUC 86.0% on Avenue, 83.9% on Ped1, 96% on Ped2 datasets. Extensive experimental results and ablation studies further validate the effectiveness of our model.
Year
DOI
Venue
2020
10.1109/TCSVT.2019.2962229
Periodicals
Keywords
DocType
Volume
Anomaly detection, Training, Task analysis, Training data, Optimization, Deep learning, Convolutional codes, Anomaly detection, deep learning, attention
Journal
30
Issue
ISSN
Citations 
12
1051-8215
9
PageRank 
References 
Authors
0.47
24
6
Name
Order
Citations
PageRank
Joey Tianyi Zhou135438.60
Le Zhang2979.97
Zhiwen Fang3142.25
Jiawei Du4171.59
xi peng5966.39
Xiao Yang6201.79