Back to Search Start Over

TEACH: Attention-Aware Deep Cross-Modal Hashing

Authors :
Yu-Wei Zhan
Xin-Shun Xu
Zhen-Duo Chen
Xin Luo
Hong-Lei Yao
Source :
ICMR
Publication Year :
2021
Publisher :
ACM, 2021.

Abstract

Hashing methods for cross-modal retrieval have recently been widely investigated due to the explosive growth of multimedia data. Generally, real-world data is imperfect and has more or less redundancy, making cross-modal retrieval task challenging. However, most existing cross-modal hashing methods fail to deal with the redundancy, leading to unsatisfactory performance on such data. In this paper, to address this issue, we propose a novel cross-modal hashing method, namely aTtEntion-Aware deep Cross-modal Hashing (TEACH). It could perform feature learning and hash-code learning simultaneously. Besides, with designed attention modules for different modalities, one for each, TEACH can effectively highlight the useful information of data while suppressing the redundant information. Extensive experiments on benchmark datasets demonstrate that our method outperforms some state-of-the-art hashing methods in cross-modal retrieval tasks.

Details

Database :
OpenAIRE
Journal :
Proceedings of the 2021 International Conference on Multimedia Retrieval
Accession number :
edsair.doi...........92b38064e603d7ef764c138f0a771d32