Title
Autoencoder-Based Self-Supervised Hashing For Cross-Modal Retrieval
Abstract
Cross-modal retrieval has gained lots of attention in the era of the multimedia data explosion. Taking advantage of low storage cost and fast retrieval speed, hash learning-based methods become more and more popular in this field. The crucial bottlenecks of cross-modal retrieval are twofold: the heterogeneous gap in different modalities and the semantic gap among similar data with various modalities. To address these issues, we adopt self-supervised fashion to bridge the heterogeneous gap by generating the cohesive features of different instances. To mitigate the semantic gap, we use triplet sampling to optimize the semantic loss in inter-modal and intra-modal, which increase the discriminability of our approach. Experimental on two benchmark datasets show the efficiency and robustness of our method, and the extended experiments show the scalability.
Year
DOI
Venue
2021
10.1007/s11042-020-09599-7
MULTIMEDIA TOOLS AND APPLICATIONS
Keywords
DocType
Volume
Cross-modal retrieval, Hash learning, Autoencoder, Self-supervised
Journal
80
Issue
ISSN
Citations 
11
1380-7501
0
PageRank 
References 
Authors
0.34
26
7
Name
Order
Citations
PageRank
Yifan Li111.16
Xuan Wang229157.12
Lizhen Cui315438.68
Zhang Jiajia436.01
Chengkai Huang510.82
Xuan Luo63010.84
Shuhan Qi73814.95