Title
Adversarial Feature Sampling Learning for Efficient Visual Tracking
Abstract
The tracking-by-detection tracking framework usually consists of two stages: drawing samples around the target object and classifying each sample as either the target object or background. Current popular trackers under this framework typically draw many samples from the raw image and feed them into the deep neural networks, resulting in high computational burden and low tracking speed. In this article, we propose an adversarial feature sampling learning (AFSL) method to address this problem. A convolutional neural network is designed, which takes only one cropped image around the target object as input, and samples are collected from the feature maps with spatial bilinear resampling. To enrich the appearance variations of positive samples in the feature space, which has limited spatial resolution, we fuse the high-level features and low-level features to better describe the target by using a generative adversarial network. Extensive experiments on benchmark data sets demonstrate that the proposed ASFL achieves leading tracking accuracy while significantly accelerating the speed of tracking-by-detection trackers. <italic xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">Note to Practitioners</italic> —Visual tracking can be applied in many intelligent automation systems, such as robotic intelligent navigation system, intelligent human–computer interaction system, and so on. In a robotic intelligent navigation system, visual tracking can generate target’s motion trajectory from image sequences. Visual tracking can also obtain body movement information automatically during the interactive process in the intelligent human–computer interaction system. Accuracy and speed are two key indicators for visual tracking, and intelligent automation systems usually need a tracker with more accuracy and faster speed. This article aims to develop a fast and accurate tracking method by adversarial feature sampling learning (AFSL). In the concrete implementation process, AFSL gets samples by sampling in the feature space rather than on raw images to reduce computation. Then, an adversarial learning mechanism is adopted to boost the sampling features and enrich the target appearances in the feature space to improve the tracking accuracy. The proposed tracker is proven to be effective to keep leading tracking accuracy while significantly accelerating the tracking speed.
Year
DOI
Venue
2020
10.1109/TASE.2019.2948402
IEEE Transactions on Automation Science and Engineering
Keywords
DocType
Volume
Adversarial learning,deep convolution neural network,feature sampling,visual tracking
Journal
17
Issue
ISSN
Citations 
2
1545-5955
1
PageRank 
References 
Authors
0.36
0
4
Name
Order
Citations
PageRank
Yingjie Yin1434.72
De Xu29010.98
Xingang Wang36910.51
Lei Zhang416326543.99