• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于大规模跨模态检索的关节细节与一致性哈希学习

Joint Specifics and Consistency Hash Learning for Large-Scale Cross-Modal Retrieval.

作者信息

Qin Jianyang, Fei Lunke, Zhang Zheng, Wen Jie, Xu Yong, Zhang David

出版信息

IEEE Trans Image Process. 2022;31:5343-5358. doi: 10.1109/TIP.2022.3195059. Epub 2022 Aug 16.

DOI:10.1109/TIP.2022.3195059
PMID:35925845
Abstract

With the dramatic increase in the amount of multimedia data, cross-modal similarity retrieval has become one of the most popular yet challenging problems. Hashing offers a promising solution for large-scale cross-modal data searching by embedding the high-dimensional data into the low-dimensional similarity preserving Hamming space. However, most existing cross-modal hashing usually seeks a semantic representation shared by multiple modalities, which cannot fully preserve and fuse the discriminative modal-specific features and heterogeneous similarity for cross-modal similarity searching. In this paper, we propose a joint specifics and consistency hash learning method for cross-modal retrieval. Specifically, we introduce an asymmetric learning framework to fully exploit the label information for discriminative hash code learning, where 1) each individual modality can be better converted into a meaningful subspace with specific information, 2) multiple subspaces are semantically connected to capture consistent information, and 3) the integration complexity of different subspaces is overcome so that the learned collaborative binary codes can merge the specifics with consistency. Then, we introduce an alternatively iterative optimization to tackle the specifics and consistency hashing learning problem, making it scalable for large-scale cross-modal retrieval. Extensive experiments on five widely used benchmark databases clearly demonstrate the effectiveness and efficiency of our proposed method on both one-cross-one and one-cross-two retrieval tasks.

摘要

随着多媒体数据量的急剧增加,跨模态相似性检索已成为最热门但也最具挑战性的问题之一。哈希通过将高维数据嵌入到低维的保持相似性的汉明空间中,为大规模跨模态数据搜索提供了一种很有前景的解决方案。然而,大多数现有的跨模态哈希通常寻求多种模态共享的语义表示,这无法充分保留和融合用于跨模态相似性搜索的判别性模态特定特征和异构相似性。在本文中,我们提出了一种用于跨模态检索的联合特异性和一致性哈希学习方法。具体来说,我们引入了一个非对称学习框架来充分利用标签信息进行判别性哈希码学习,其中:1)每个单独的模态可以更好地转换为具有特定信息的有意义子空间;2)多个子空间在语义上相互连接以捕获一致信息;3)克服了不同子空间的集成复杂性,以便学习到的协作二进制码能够将特异性与一致性融合起来。然后,我们引入了一种交替迭代优化方法来解决特异性和一致性哈希学习问题,使其能够扩展到大规模跨模态检索。在五个广泛使用的基准数据库上进行的大量实验清楚地证明了我们提出的方法在单对单和单对二检索任务上的有效性和效率。

相似文献

1
Joint Specifics and Consistency Hash Learning for Large-Scale Cross-Modal Retrieval.用于大规模跨模态检索的关节细节与一致性哈希学习
IEEE Trans Image Process. 2022;31:5343-5358. doi: 10.1109/TIP.2022.3195059. Epub 2022 Aug 16.
2
Multimodal Discriminative Binary Embedding for Large-Scale Cross-Modal Retrieval.多模态判别式二值嵌入的大规模跨模态检索。
IEEE Trans Image Process. 2016 Oct;25(10):4540-54. doi: 10.1109/TIP.2016.2592800. Epub 2016 Jul 18.
3
Linear Subspace Ranking Hashing for Cross-Modal Retrieval.线性子空间排序哈希用于跨模态检索。
IEEE Trans Pattern Anal Mach Intell. 2017 Sep;39(9):1825-1838. doi: 10.1109/TPAMI.2016.2610969. Epub 2016 Sep 19.
4
Structure-aware contrastive hashing for unsupervised cross-modal retrieval.用于无监督跨模态检索的结构感知对比哈希
Neural Netw. 2024 Jun;174:106211. doi: 10.1016/j.neunet.2024.106211. Epub 2024 Feb 27.
5
FDDH: Fast Discriminative Discrete Hashing for Large-Scale Cross-Modal Retrieval.FDDH:用于大规模跨模态检索的快速判别离散哈希算法
IEEE Trans Neural Netw Learn Syst. 2022 Nov;33(11):6306-6320. doi: 10.1109/TNNLS.2021.3076684. Epub 2022 Oct 27.
6
Deep Semantic-Preserving Ordinal Hashing for Cross-Modal Similarity Search.用于跨模态相似性搜索的深度语义保持序数哈希
IEEE Trans Neural Netw Learn Syst. 2019 May;30(5):1429-1440. doi: 10.1109/TNNLS.2018.2869601. Epub 2018 Oct 1.
7
Learning Discriminative Binary Codes for Large-scale Cross-modal Retrieval.学习判别式二进制代码进行大规模跨模态检索。
IEEE Trans Image Process. 2017 May;26(5):2494-2507. doi: 10.1109/TIP.2017.2676345. Epub 2017 Mar 1.
8
Semantic embedding based online cross-modal hashing method.基于语义嵌入的在线跨模态哈希方法。
Sci Rep. 2024 Jan 6;14(1):736. doi: 10.1038/s41598-023-50242-w.
9
Deep Semantic Multimodal Hashing Network for Scalable Image-Text and Video-Text Retrievals.用于可扩展图像-文本和视频-文本检索的深度语义多模态哈希网络
IEEE Trans Neural Netw Learn Syst. 2023 Apr;34(4):1838-1851. doi: 10.1109/TNNLS.2020.2997020. Epub 2023 Apr 4.
10
Triplet-Based Deep Hashing Network for Cross-Modal Retrieval.用于跨模态检索的基于三元组的深度哈希网络。
IEEE Trans Image Process. 2018 Aug;27(8):3893-3903. doi: 10.1109/TIP.2018.2821921. Epub 2018 Apr 4.