TY - JOUR
T1 - Deep supervised fused similarity hashing for cross-modal retrieval
AU - Ng, Wing
AU - Xu, Yongzhi
AU - Tian, Xing
AU - Wang, Hui
PY - 2024/6/21
Y1 - 2024/6/21
N2 - The need for cross-modal retrieval increases significantly with the rapid growth of multimedia information on the Internet. However, most of existing cross-modal retrieval methods neglect the correlation between label similarity and intra-modality similarity in common semantic subspace training, which makes the trained common semantic subspace unable to preserve semantic similarity of original data effectively. Therefore, a novel cross-modal hashing method is proposed in this paper, namely, Deep Supervised Fused Similarity Hashing (DSFSH). The DSFSH mainly consists of two parts. Firstly, a fused similarity method is proposed to exploit the intrinsic inter-modality correlation of data while preserving the intra-modality relationship of data at the same time. Secondly, a novel quantization max-margin loss is proposed. The gap between cosine similarity and Hamming similarity is closed by minimizing this loss. Extensive experimental results on three benchmark datasets show that the proposed method yields better retrieval performance comparing to state-of-the-art methods.
AB - The need for cross-modal retrieval increases significantly with the rapid growth of multimedia information on the Internet. However, most of existing cross-modal retrieval methods neglect the correlation between label similarity and intra-modality similarity in common semantic subspace training, which makes the trained common semantic subspace unable to preserve semantic similarity of original data effectively. Therefore, a novel cross-modal hashing method is proposed in this paper, namely, Deep Supervised Fused Similarity Hashing (DSFSH). The DSFSH mainly consists of two parts. Firstly, a fused similarity method is proposed to exploit the intrinsic inter-modality correlation of data while preserving the intra-modality relationship of data at the same time. Secondly, a novel quantization max-margin loss is proposed. The gap between cosine similarity and Hamming similarity is closed by minimizing this loss. Extensive experimental results on three benchmark datasets show that the proposed method yields better retrieval performance comparing to state-of-the-art methods.
U2 - 10.1007/s11042-024-19581-2
DO - 10.1007/s11042-024-19581-2
M3 - Article
SN - 1380-7501
JO - Multimedia Tools and Applications
JF - Multimedia Tools and Applications
ER -