Title
Deep Supervised Cross-Modal Retrieval
Abstract
Cross-modal retrieval aims to enable flexible retrieval across different modalities. The core of cross-modal retrieval is how to measure the content similarity between different types of data. In this paper, we present a novel cross-modal retrieval method, called Deep Supervised Cross-modal Retrieval (DSCMR). It aims to find a common representation space, in which the samples from different modalities can be compared directly. Specifically, DSCMR minimises the discrimination loss in both the label space and the common representation space to supervise the model learning discriminative features. Furthermore, it simultaneously minimises the modality invariance loss and uses a weight sharing strategy to eliminate the cross-modal discrepancy of multimedia data in the common representation space to learn modality-invariant features. Comprehensive experimental results on four widely-used benchmark datasets demonstrate that the proposed method is effective in cross-modal learning and significantly outperforms the state-of-the-art cross-modal retrieval methods.
Year
DOI
Venue
2019
10.1109/CVPR.2019.01064
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019)
Field
DocType
ISSN
Pattern recognition,Computer science,Artificial intelligence,Modal
Conference
1063-6919
Citations 
PageRank 
References 
11
0.47
0
Authors
4
Name
Order
Citations
PageRank
Liangli Zhen1729.73
Peng Hu2719.06
Xu Wang310315.76
Dezhong Peng428527.92