• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于无监督跨模态检索的深度语义保持重构哈希

Deep Semantic-Preserving Reconstruction Hashing for Unsupervised Cross-Modal Retrieval.

作者信息

Cheng Shuli, Wang Liejun, Du Anyu

机构信息

College of Information Science and Engineering, Xinjiang University, Urumqi 830046, China.

Key Laboratory of Signal Detection and Processing, Xinjiang Uygur Autonomous Region, Xinjiang University, Urumqi 830046, China.

出版信息

Entropy (Basel). 2020 Nov 7;22(11):1266. doi: 10.3390/e22111266.

DOI:10.3390/e22111266
PMID:33287034
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7712897/
Abstract

Deep hashing is the mainstream algorithm for large-scale cross-modal retrieval due to its high retrieval speed and low storage capacity, but the problem of reconstruction of modal semantic information is still very challenging. In order to further solve the problem of unsupervised cross-modal retrieval semantic reconstruction, we propose a novel deep semantic-preserving reconstruction hashing (DSPRH). The algorithm combines spatial and channel semantic information, and mines modal semantic information based on adaptive self-encoding and joint semantic reconstruction loss. The main contributions are as follows: (1) We introduce a new spatial pooling network module based on tensor regular-polymorphic decomposition theory to generate rank-1 tensor to capture high-order context semantics, which can assist the backbone network to capture important contextual modal semantic information. (2) Based on optimization perspective, we use global covariance pooling to capture channel semantic information and accelerate network convergence. In feature reconstruction layer, we use two bottlenecks auto-encoding to achieve visual-text modal interaction. (3) In metric learning, we design a new loss function to optimize model parameters, which can preserve the correlation between image modalities and text modalities. The DSPRH algorithm is tested on MIRFlickr-25K and NUS-WIDE. The experimental results show that DSPRH has achieved better performance on retrieval tasks.

摘要

深度哈希由于其检索速度快和存储容量低,是大规模跨模态检索的主流算法,但模态语义信息的重建问题仍然非常具有挑战性。为了进一步解决无监督跨模态检索语义重建问题,我们提出了一种新颖的深度语义保持重建哈希(DSPRH)算法。该算法结合了空间和通道语义信息,并基于自适应自编码和联合语义重建损失挖掘模态语义信息。主要贡献如下:(1)我们引入了一种基于张量正则多态分解理论的新的空间池化网络模块,以生成秩-1张量来捕获高阶上下文语义,这可以帮助骨干网络捕获重要的上下文模态语义信息。(2)从优化的角度出发,我们使用全局协方差池化来捕获通道语义信息并加速网络收敛。在特征重建层,我们使用两个瓶颈自编码来实现视觉-文本模态交互。(3)在度量学习中,我们设计了一种新的损失函数来优化模型参数,该函数可以保持图像模态和文本模态之间的相关性。DSPRH算法在MIRFlickr-25K和NUS-WIDE上进行了测试。实验结果表明,DSPRH在检索任务上取得了更好的性能。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6b53/7712897/418ec7fee559/entropy-22-01266-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6b53/7712897/a1e0cd1733dc/entropy-22-01266-g0A1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6b53/7712897/6b17391833b8/entropy-22-01266-g0A2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6b53/7712897/7e792bfa2bb4/entropy-22-01266-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6b53/7712897/809ab5a164a3/entropy-22-01266-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6b53/7712897/784678671997/entropy-22-01266-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6b53/7712897/44e513bf126d/entropy-22-01266-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6b53/7712897/d08b20201cc6/entropy-22-01266-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6b53/7712897/418ec7fee559/entropy-22-01266-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6b53/7712897/a1e0cd1733dc/entropy-22-01266-g0A1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6b53/7712897/6b17391833b8/entropy-22-01266-g0A2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6b53/7712897/7e792bfa2bb4/entropy-22-01266-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6b53/7712897/809ab5a164a3/entropy-22-01266-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6b53/7712897/784678671997/entropy-22-01266-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6b53/7712897/44e513bf126d/entropy-22-01266-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6b53/7712897/d08b20201cc6/entropy-22-01266-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6b53/7712897/418ec7fee559/entropy-22-01266-g006.jpg

相似文献

1
Deep Semantic-Preserving Reconstruction Hashing for Unsupervised Cross-Modal Retrieval.用于无监督跨模态检索的深度语义保持重构哈希
Entropy (Basel). 2020 Nov 7;22(11):1266. doi: 10.3390/e22111266.
2
Deep Unsupervised Hashing for Large-Scale Cross-Modal Retrieval Using Knowledge Distillation Model.基于知识蒸馏模型的大规模跨模态检索的深度无监督哈希。
Comput Intell Neurosci. 2021 Jul 17;2021:5107034. doi: 10.1155/2021/5107034. eCollection 2021.
3
Structure-aware contrastive hashing for unsupervised cross-modal retrieval.用于无监督跨模态检索的结构感知对比哈希
Neural Netw. 2024 Jun;174:106211. doi: 10.1016/j.neunet.2024.106211. Epub 2024 Feb 27.
4
Deep Semantic Multimodal Hashing Network for Scalable Image-Text and Video-Text Retrievals.用于可扩展图像-文本和视频-文本检索的深度语义多模态哈希网络
IEEE Trans Neural Netw Learn Syst. 2023 Apr;34(4):1838-1851. doi: 10.1109/TNNLS.2020.2997020. Epub 2023 Apr 4.
5
Hierarchical semantic interaction-based deep hashing network for cross-modal retrieval.基于层次语义交互的深度哈希网络用于跨模态检索。
PeerJ Comput Sci. 2021 May 25;7:e552. doi: 10.7717/peerj-cs.552. eCollection 2021.
6
Quadruplet-Based Deep Cross-Modal Hashing.四元组深度学习跨模态哈希。
Comput Intell Neurosci. 2021 Jul 2;2021:9968716. doi: 10.1155/2021/9968716. eCollection 2021.
7
Object-Level Visual-Text Correlation Graph Hashing for Unsupervised Cross-Modal Retrieval.用于无监督跨模态检索的对象级视觉-文本关联图哈希
Sensors (Basel). 2022 Apr 11;22(8):2921. doi: 10.3390/s22082921.
8
CLIP-Based Adaptive Graph Attention Network for Large-Scale Unsupervised Multi-Modal Hashing Retrieval.基于 CLIP 的自适应图注意力网络的大规模无监督多模态哈希检索
Sensors (Basel). 2023 Mar 24;23(7):3439. doi: 10.3390/s23073439.
9
Hierarchical Recurrent Neural Hashing for Image Retrieval With Hierarchical Convolutional Features.基于层次卷积特征的层次递归神经网络哈希图像检索
IEEE Trans Image Process. 2018;27(1):106-120. doi: 10.1109/TIP.2017.2755766.
10
Unsupervised Dual Deep Hashing With Semantic-Index and Content-Code for Cross-Modal Retrieval.基于语义索引和内容编码的无监督双深度哈希用于跨模态检索
IEEE Trans Pattern Anal Mach Intell. 2025 Jan;47(1):387-399. doi: 10.1109/TPAMI.2024.3467130. Epub 2024 Dec 4.

引用本文的文献

1
Deep parameter-free attention hashing for image retrieval.深度无参注意力哈希检索图像。
Sci Rep. 2022 Apr 30;12(1):7082. doi: 10.1038/s41598-022-11217-5.

本文引用的文献

1
Deep CNNs Meet Global Covariance Pooling: Better Representation and Generalization.深度卷积神经网络与全局协方差池化:更好的表示和泛化。
IEEE Trans Pattern Anal Mach Intell. 2021 Aug;43(8):2582-2597. doi: 10.1109/TPAMI.2020.2974833. Epub 2021 Jul 1.
2
Multi-Task Consistency-Preserving Adversarial Hashing for Cross-Modal Retrieval.用于跨模态检索的多任务一致性保持对抗哈希
IEEE Trans Image Process. 2020 Jan 9. doi: 10.1109/TIP.2020.2963957.
3
MTFH: A Matrix Tri-Factorization Hashing Framework for Efficient Cross-Modal Retrieval.
MTFH:一种用于高效跨模态检索的矩阵三因素分解哈希框架。
IEEE Trans Pattern Anal Mach Intell. 2021 Mar;43(3):964-981. doi: 10.1109/TPAMI.2019.2940446. Epub 2021 Feb 4.
4
Attribute-Guided Network for Cross-Modal Zero-Shot Hashing.用于跨模态零样本哈希的属性引导网络。
IEEE Trans Neural Netw Learn Syst. 2020 Jan;31(1):321-330. doi: 10.1109/TNNLS.2019.2904991. Epub 2019 Apr 11.
5
Collective Reconstructive Embeddings for Cross-modal Hashing.用于跨模态哈希的集体重构嵌入
IEEE Trans Image Process. 2018 Dec 28. doi: 10.1109/TIP.2018.2890144.
6
Deep Semantic-Preserving Ordinal Hashing for Cross-Modal Similarity Search.用于跨模态相似性搜索的深度语义保持序数哈希
IEEE Trans Neural Netw Learn Syst. 2019 May;30(5):1429-1440. doi: 10.1109/TNNLS.2018.2869601. Epub 2018 Oct 1.
7
SCH-GAN: Semi-Supervised Cross-Modal Hashing by Generative Adversarial Network.SCH-GAN:基于生成对抗网络的半监督跨模态哈希。
IEEE Trans Cybern. 2020 Feb;50(2):489-502. doi: 10.1109/TCYB.2018.2868826. Epub 2018 Sep 26.