DSCEH: Dual-Stream Correlation-Enhanced Deep Hashing for Image Retrieval

Author:

Yang Yulin1,Chen Huizhen1,Liu Rongkai1,Liu Shuning1,Zhan Yu2,Hu Chao3,Shi Ronghua3

Affiliation:

1. School of Computer Science and Engineering, Central South University, Changsha 410083, China

2. China Telecom, Changsha 410083, China

3. School of Electronic Information, Central South University, Changsha 410083, China

Abstract

Deep Hashing is widely used for large-scale image-retrieval tasks to speed up the retrieval process. Current deep hashing methods are mainly based on the Convolutional Neural Network (CNN) or Vision Transformer (VIT). They only use the local or global features for low-dimensional mapping and only use the similarity loss function to optimize the correlation between pairwise or triplet images. Therefore, the effectiveness of deep hashing methods is limited. In this paper, we propose a dual-stream correlation-enhanced deep hashing framework (DSCEH), which uses the local and global features of the image for low-dimensional mapping and optimizes the correlation of images from the model architecture. DSCEH consists of two main steps: model training and deep-hash-based retrieval. During the training phase, a dual-network structure comprising CNN and VIT is employed for feature extraction. Subsequently, feature fusion is achieved through a concatenation operation, followed by similarity evaluation based on the class token acquired from VIT to establish edge relationships. The Graph Convolutional Network is then utilized to enhance correlation optimization between images, resulting in the generation of high-quality hash codes. This stage facilitates the development of an optimized hash model for image retrieval. In the retrieval stage, all images within the database and the to-be-retrieved images are initially mapped to hash codes using the aforementioned hash model. The retrieval results are subsequently determined based on the Hamming distance between the hash codes. We conduct experiments on three datasets: CIFAR-10, MSCOCO, and NUSWIDE. Experimental results show the superior performance of DSCEH, which helps with fast and accurate image retrieval.

Funder

National Natural Science Foundation of China

Hunan Provincial Educational Science Research Base Project

Philosophy and Social Sciences Foundation of Hunan Province

Hunan Province Science and Technology Innovation Project

High Performance Computing Center of Central South University

Publisher

MDPI AG

Reference43 articles.

1. Oria, V., Sapino, M.L., Satoh, S., Kerhervé, B., Cheng, W., Ide, I., and Singh, V.K. (2022, January 27–30). TransHash: Transformer-based Hamming Hashing for Efficient Image Retrieval. Proceedings of the ICMR ’22: International Conference on Multimedia Retrieval, Newark, NJ, USA.

2. Jang, J., Choi, H., Bae, H., Lee, S., Kwon, M., and Jung, M. (2023, January 10–12). CXL-ANNS: Software-Hardware Collaborative Memory Disaggregation and Computation for Billion-Scale Approximate Nearest Neighbor Search. Proceedings of the USENIX Annual Technical Conference, Boston, MA, USA.

3. Query-Adaptive Image Retrieval by Deep-Weighted Hashing;Zhang;IEEE Trans. Multim.,2018

4. Scalable Discrete and Asymmetric Unequal Length Hashing Learning for Cross-Modal Retrieval;Teng;IEEE Trans. Multim.,2024

5. Deep global semantic structure-preserving hashing via corrective triplet loss for remote sensing image retrieval;Zhou;Expert Syst. Appl.,2024

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3