Title
Learning reliable modal weight with transformer for robust RGBT tracking
Abstract
Many Siamese-based RGBT trackers have been prevalently designed in recent years for fast-tracking. However, the correlation operation in them is a local linear matching process, which may easily lose semantic information required inevitably by high-precision trackers. In this paper, we propose a strong cross-modal model based on transformer for robust RGBT tracking. Specifically, a simple dual-flow convolutional network is designed to extract and fuse dual-modal features, with comparably lower complexity. Besides, to enhance the feature representation and deepen semantic features, a modal weight allocation strategy and a backbone feature extracted network based on modified Resnet-50 are designed, respectively. Also, an attention-based transformer feature fusion network is adopted to improve long-distance feature association to decrease the loss of semantic information. Finally, a classification regression subnetwork is investigated to accurately predict the state of the target. Sufficient experiments have been implemented on the RGBT234, RGBT210, GTOT and LasHeR datasets, demonstrating more outstanding tracking performance against the state-of-the-art RGBT trackers.
Year
DOI
Venue
2022
10.1016/j.knosys.2022.108945
Knowledge-Based Systems
Keywords
DocType
Volume
RGBT tracking,Transformer,Semantic features
Journal
249
ISSN
Citations 
PageRank 
0950-7051
0
0.34
References 
Authors
0
2
Name
Order
Citations
PageRank
Mingzheng Feng100.34
Jianbo Su223138.20