151. Binary Set Embedding for Cross-Modal Retrieval
- Author
-
Mengyang Yu, Li Liu, and Ling Shao
- Subjects
Computer Networks and Communications ,Computer science ,business.industry ,Hash function ,Feature extraction ,Pattern recognition ,02 engineering and technology ,010501 environmental sciences ,01 natural sciences ,Computer Science Applications ,Image (mathematics) ,Artificial Intelligence ,Feature (computer vision) ,0202 electrical engineering, electronic engineering, information engineering ,Embedding ,020201 artificial intelligence & image processing ,Binary code ,Visual Word ,Artificial intelligence ,Hamming space ,business ,Orthogonalization ,Software ,0105 earth and related environmental sciences ,Semantic gap - Abstract
Cross-modal retrieval is such a challenging topic that traditional global representations would fail to bridge the semantic gap between images and texts to a satisfactory level. Using local features from images and words from documents directly can be more robust for the scenario with large intraclass variations and small interclass discrepancies. In this paper, we propose a novel unsupervised binary coding algorithm called binary set embedding (BSE) to obtain meaningful hash codes for local features from the image domain and words from text domain. Understanding image features with the word vectors learned from the human language instead of the provided documents from data sets, BSE can map samples into a common Hamming space effectively and efficiently where each sample is represented by the sets of local feature descriptors from image and text domains. In particular, BSE explores relationship among local features in both feature level and image (text) level, which can balance the sensitivity of each other. Furthermore, a recursive orthogonalization procedure is applied to reduce the redundancy of codes. Extensive experiments demonstrate the superior performance of BSE compared with state-of-the-art cross-modal hashing methods using either image or text queries.
- Published
- 2017