Back to Search
Start Over
TEACH: Attention-Aware Deep Cross-Modal Hashing
- Source :
- ICMR
- Publication Year :
- 2021
- Publisher :
- ACM, 2021.
-
Abstract
- Hashing methods for cross-modal retrieval have recently been widely investigated due to the explosive growth of multimedia data. Generally, real-world data is imperfect and has more or less redundancy, making cross-modal retrieval task challenging. However, most existing cross-modal hashing methods fail to deal with the redundancy, leading to unsatisfactory performance on such data. In this paper, to address this issue, we propose a novel cross-modal hashing method, namely aTtEntion-Aware deep Cross-modal Hashing (TEACH). It could perform feature learning and hash-code learning simultaneously. Besides, with designed attention modules for different modalities, one for each, TEACH can effectively highlight the useful information of data while suppressing the redundant information. Extensive experiments on benchmark datasets demonstrate that our method outperforms some state-of-the-art hashing methods in cross-modal retrieval tasks.
Details
- Database :
- OpenAIRE
- Journal :
- Proceedings of the 2021 International Conference on Multimedia Retrieval
- Accession number :
- edsair.doi...........92b38064e603d7ef764c138f0a771d32