Abstract | ||
---|---|---|
AbstractRecent video anomaly detection methods focus on reconstructing or predicting frames. Under this umbrella, the long-standing inter-class data-imbalance problem resorts to the imbalance between foreground and stationary background objects in video anomaly detection and this has been less investigated by existing solutions. Naively optimizing the reconstructing loss yields a biased optimization towards background reconstruction rather than the objects of interest in the foreground. To solve this, we proposed a simple yet effective solution, termed attention-driven loss to alleviate the foreground-background imbalance problem in anomaly detection. Specifically, we compute a single mask map that summarizes the frame evolution of moving foreground regions and suppresses the background in the training video clips. After that, we construct an attention map through the combination of the mask map and background to give different weights to the foreground and background region respectively. The proposed attention-driven loss is independent of backbone networks and can be easily augmented in most existing anomaly detection models. Augmented with attention-driven loss, the model is able to achieve AUC 86.0% on Avenue, 83.9% on Ped1, 96% on Ped2 datasets. Extensive experimental results and ablation studies further validate the effectiveness of our model. |
Year | DOI | Venue |
---|---|---|
2020 | 10.1109/TCSVT.2019.2962229 | Periodicals |
Keywords | DocType | Volume |
Anomaly detection, Training, Task analysis, Training data, Optimization, Deep learning, Convolutional codes, Anomaly detection, deep learning, attention | Journal | 30 |
Issue | ISSN | Citations |
12 | 1051-8215 | 9 |
PageRank | References | Authors |
0.47 | 24 | 6 |
Name | Order | Citations | PageRank |
---|---|---|---|
Joey Tianyi Zhou | 1 | 354 | 38.60 |
Le Zhang | 2 | 97 | 9.97 |
Zhiwen Fang | 3 | 14 | 2.25 |
Jiawei Du | 4 | 17 | 1.59 |
xi peng | 5 | 96 | 6.39 |
Xiao Yang | 6 | 20 | 1.79 |