• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

共享预测跨模态深度量化

Shared Predictive Cross-Modal Deep Quantization.

作者信息

Yang Erkun, Deng Cheng, Li Chao, Liu Wei, Li Jie, Tao Dacheng

出版信息

IEEE Trans Neural Netw Learn Syst. 2018 Nov;29(11):5292-5303. doi: 10.1109/TNNLS.2018.2793863. Epub 2018 Feb 14.

DOI:10.1109/TNNLS.2018.2793863
PMID:29994640
Abstract

With explosive growth of data volume and ever-increasing diversity of data modalities, cross-modal similarity search, which conducts nearest neighbor search across different modalities, has been attracting increasing interest. This paper presents a deep compact code learning solution for efficient cross-modal similarity search. Many recent studies have proven that quantization-based approaches perform generally better than hashing-based approaches on single-modal similarity search. In this paper, we propose a deep quantization approach, which is among the early attempts of leveraging deep neural networks into quantization-based cross-modal similarity search. Our approach, dubbed shared predictive deep quantization (SPDQ), explicitly formulates a shared subspace across different modalities and two private subspaces for individual modalities, and representations in the shared subspace and the private subspaces are learned simultaneously by embedding them to a reproducing kernel Hilbert space, where the mean embedding of different modality distributions can be explicitly compared. In addition, in the shared subspace, a quantizer is learned to produce the semantics preserving compact codes with the help of label alignment. Thanks to this novel network architecture in cooperation with supervised quantization training, SPDQ can preserve intramodal and intermodal similarities as much as possible and greatly reduce quantization error. Experiments on two popular benchmarks corroborate that our approach outperforms state-of-the-art methods.

摘要

随着数据量的爆炸式增长以及数据模态的日益多样化,跨模态相似性搜索(即在不同模态间进行最近邻搜索)已引起越来越多的关注。本文提出了一种用于高效跨模态相似性搜索的深度紧凑编码学习解决方案。最近的许多研究表明,在单模态相似性搜索中,基于量化的方法通常比基于哈希的方法表现更好。在本文中,我们提出了一种深度量化方法,这是将深度神经网络应用于基于量化的跨模态相似性搜索的早期尝试之一。我们的方法称为共享预测深度量化(SPDQ),它明确地在不同模态间构建一个共享子空间以及为各个模态构建两个私有子空间,并通过将它们嵌入到再生核希尔伯特空间中同时学习共享子空间和私有子空间中的表示,在该空间中可以明确比较不同模态分布的均值嵌入。此外,在共享子空间中,借助标签对齐学习一个量化器以生成保留语义的紧凑编码。得益于这种新颖的网络架构与监督量化训练的结合,SPDQ能够尽可能多地保留模态内和模态间的相似性,并大大降低量化误差。在两个流行基准上的实验证实,我们的方法优于现有方法。

相似文献

1
Shared Predictive Cross-Modal Deep Quantization.共享预测跨模态深度量化
IEEE Trans Neural Netw Learn Syst. 2018 Nov;29(11):5292-5303. doi: 10.1109/TNNLS.2018.2793863. Epub 2018 Feb 14.
2
Multimodal Discriminative Binary Embedding for Large-Scale Cross-Modal Retrieval.多模态判别式二值嵌入的大规模跨模态检索。
IEEE Trans Image Process. 2016 Oct;25(10):4540-54. doi: 10.1109/TIP.2016.2592800. Epub 2016 Jul 18.
3
Deep Semantic-Preserving Ordinal Hashing for Cross-Modal Similarity Search.用于跨模态相似性搜索的深度语义保持序数哈希
IEEE Trans Neural Netw Learn Syst. 2019 May;30(5):1429-1440. doi: 10.1109/TNNLS.2018.2869601. Epub 2018 Oct 1.
4
Joint Specifics and Consistency Hash Learning for Large-Scale Cross-Modal Retrieval.用于大规模跨模态检索的关节细节与一致性哈希学习
IEEE Trans Image Process. 2022;31:5343-5358. doi: 10.1109/TIP.2022.3195059. Epub 2022 Aug 16.
5
Sequential Discrete Hashing for Scalable Cross-Modality Similarity Retrieval.用于可扩展跨模态相似性检索的序贯离散哈希。
IEEE Trans Image Process. 2017 Jan;26(1):107-118. doi: 10.1109/TIP.2016.2619262. Epub 2016 Oct 19.
6
Structure Sensitive Hashing With Adaptive Product Quantization.结构敏感哈希与自适应乘积量化。
IEEE Trans Cybern. 2016 Oct;46(10):2252-2264. doi: 10.1109/TCYB.2015.2474742. Epub 2015 Oct 1.
7
Learning Cross-Modal Common Representations by Private-Shared Subspaces Separation.通过私有共享子空间分离学习跨模态公共表示。
IEEE Trans Cybern. 2022 May;52(5):3261-3275. doi: 10.1109/TCYB.2020.3009004. Epub 2022 May 19.
8
Large-Scale Unsupervised Hashing with Shared Structure Learning.大规模无监督哈希共享结构学习。
IEEE Trans Cybern. 2015 Sep;45(9):1811-22. doi: 10.1109/TCYB.2014.2360856. Epub 2014 Nov 20.
9
Multimodal Mutual Information Maximization: A Novel Approach for Unsupervised Deep Cross-Modal Hashing.多模态互信息最大化:一种用于无监督深度跨模态哈希的新方法。
IEEE Trans Neural Netw Learn Syst. 2023 Sep;34(9):6289-6302. doi: 10.1109/TNNLS.2021.3135420. Epub 2023 Sep 1.
10
Learning Discriminative Binary Codes for Large-scale Cross-modal Retrieval.学习判别式二进制代码进行大规模跨模态检索。
IEEE Trans Image Process. 2017 May;26(5):2494-2507. doi: 10.1109/TIP.2017.2676345. Epub 2017 Mar 1.

引用本文的文献

1
Deep Bayesian Quantization for Supervised Neuroimage Search.用于监督神经图像搜索的深度贝叶斯量化
Mach Learn Med Imaging. 2023 Oct;14349:396-406. doi: 10.1007/978-3-031-45676-3_40. Epub 2023 Oct 15.
2
Deep Disentangled Hashing with Momentum Triplets for Neuroimage Search.用于神经图像搜索的带动量三元组的深度解缠哈希
Med Image Comput Comput Assist Interv. 2020;12261:191-201. doi: 10.1007/978-3-030-59710-8_19. Epub 2020 Sep 29.
3
Quadruplet-Based Deep Cross-Modal Hashing.四元组深度学习跨模态哈希。
Comput Intell Neurosci. 2021 Jul 2;2021:9968716. doi: 10.1155/2021/9968716. eCollection 2021.
4
A Deep Spatial Context Guided Framework for Infant Brain Subcortical Segmentation.一种用于婴儿脑皮质下分割的深度空间上下文引导框架。
Med Image Comput Comput Assist Interv. 2020 Oct;12267:646-656. doi: 10.1007/978-3-030-59728-3_63. Epub 2020 Sep 29.
5
A Mutual Multi-Scale Triplet Graph Convolutional Network for Classification of Brain Disorders Using Functional or Structural Connectivity.基于功能或结构连接的脑疾病分类的互多尺度三重图卷积网络。
IEEE Trans Med Imaging. 2021 Apr;40(4):1279-1289. doi: 10.1109/TMI.2021.3051604. Epub 2021 Apr 1.
6
Cross-Modal Search for Social Networks via Adversarial Learning.基于对抗学习的社交网络跨模态检索。
Comput Intell Neurosci. 2020 Jul 11;2020:7834953. doi: 10.1155/2020/7834953. eCollection 2020.