• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

使用混合深度信念网络学习文档语义表示。

Learning document semantic representation with hybrid deep belief network.

作者信息

Yan Yan, Yin Xu-Cheng, Li Sujian, Yang Mingyuan, Hao Hong-Wei

机构信息

Department of Computer Science and Technology, School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China.

Key Laboratory of Computational Linguistics, Peking University, Ministry of Education, Beijing 100871, China.

出版信息

Comput Intell Neurosci. 2015;2015:650527. doi: 10.1155/2015/650527. Epub 2015 Mar 23.

DOI:10.1155/2015/650527
PMID:25878657
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC4386712/
Abstract

High-level abstraction, for example, semantic representation, is vital for document classification and retrieval. However, how to learn document semantic representation is still a topic open for discussion in information retrieval and natural language processing. In this paper, we propose a new Hybrid Deep Belief Network (HDBN) which uses Deep Boltzmann Machine (DBM) on the lower layers together with Deep Belief Network (DBN) on the upper layers. The advantage of DBM is that it employs undirected connection when training weight parameters which can be used to sample the states of nodes on each layer more successfully and it is also an effective way to remove noise from the different document representation type; the DBN can enhance extract abstract of the document in depth, making the model learn sufficient semantic representation. At the same time, we explore different input strategies for semantic distributed representation. Experimental results show that our model using the word embedding instead of single word has better performance.

摘要

例如,高级抽象(即语义表示)对于文档分类和检索至关重要。然而,如何学习文档语义表示仍是信息检索和自然语言处理中一个有待探讨的话题。在本文中,我们提出了一种新的混合深度信念网络(HDBN),它在较低层使用深度玻尔兹曼机(DBM),在上层使用深度信念网络(DBN)。DBM的优势在于,在训练权重参数时采用无向连接,这能够更成功地用于对每层节点的状态进行采样,并且这也是从不同文档表示类型中去除噪声的有效方法;DBN可以深入增强对文档摘要的提取,使模型学习到足够的语义表示。同时,我们探索了语义分布式表示的不同输入策略。实验结果表明,我们使用词嵌入而非单个单词的模型具有更好的性能。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/37d1/4386712/49f1912ecede/CIN2015-650527.004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/37d1/4386712/e38c404c08af/CIN2015-650527.001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/37d1/4386712/71a8d28f142b/CIN2015-650527.002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/37d1/4386712/99986c91d00d/CIN2015-650527.003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/37d1/4386712/49f1912ecede/CIN2015-650527.004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/37d1/4386712/e38c404c08af/CIN2015-650527.001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/37d1/4386712/71a8d28f142b/CIN2015-650527.002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/37d1/4386712/99986c91d00d/CIN2015-650527.003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/37d1/4386712/49f1912ecede/CIN2015-650527.004.jpg

相似文献

1
Learning document semantic representation with hybrid deep belief network.使用混合深度信念网络学习文档语义表示。
Comput Intell Neurosci. 2015;2015:650527. doi: 10.1155/2015/650527. Epub 2015 Mar 23.
2
Biomedical literature classification with a CNNs-based hybrid learning network.基于 CNNs 的混合学习网络的生物医学文献分类。
PLoS One. 2018 Jul 26;13(7):e0197933. doi: 10.1371/journal.pone.0197933. eCollection 2018.
3
Event Recognition Based on Deep Learning in Chinese Texts.基于深度学习的中文文本事件识别
PLoS One. 2016 Aug 8;11(8):e0160147. doi: 10.1371/journal.pone.0160147. eCollection 2016.
4
Multidimensional latent semantic analysis using term spatial information.多维潜在语义分析利用术语空间信息。
IEEE Trans Cybern. 2013 Dec;43(6):1625-40. doi: 10.1109/TSMCC.2012.2227112.
5
Biomedical event trigger detection by dependency-based word embedding.基于依存关系的词嵌入进行生物医学事件触发检测
BMC Med Genomics. 2016 Aug 10;9 Suppl 2(Suppl 2):45. doi: 10.1186/s12920-016-0203-8.
6
Discovering binary codes for documents by learning deep generative models.通过学习深度生成模型发现文档的二进制代码。
Top Cogn Sci. 2011 Jan;3(1):74-91. doi: 10.1111/j.1756-8765.2010.01109.x. Epub 2010 Aug 18.
7
Novel deep generative simultaneous recurrent model for efficient representation learning.新型深度生成式同时递归模型,用于高效的表示学习。
Neural Netw. 2018 Nov;107:12-22. doi: 10.1016/j.neunet.2018.04.020. Epub 2018 Aug 9.
8
Enhancing text categorization with semantic-enriched representation and training data augmentation.通过语义丰富的表示和训练数据增强来提升文本分类
J Am Med Inform Assoc. 2006 Sep-Oct;13(5):526-35. doi: 10.1197/jamia.M2051. Epub 2006 Jun 23.
9
Visual Exploration of Neural Document Embedding in Information Retrieval: Semantics and Feature Selection.信息检索中神经文档嵌入的可视化探索:语义和特征选择。
IEEE Trans Vis Comput Graph. 2019 Jun;25(6):2181-2192. doi: 10.1109/TVCG.2019.2903946. Epub 2019 Mar 15.
10
ECRL: an eligibility criteria representation language based on the UMLS Semantic Network.ECRL:一种基于统一医学语言系统语义网络的合格标准表示语言。
AMIA Annu Symp Proc. 2008 Nov 6:1084.

引用本文的文献

1
Gesture Recognition Based on Multiscale Singular Value Entropy and Deep Belief Network.基于多尺度奇异值摘和深度置信网络的手势识别。
Sensors (Basel). 2020 Dec 27;21(1):119. doi: 10.3390/s21010119.
2
A Neural Network Approach to Intention Modeling for User-Adapted Conversational Agents.一种用于用户自适应对话代理的意图建模的神经网络方法。
Comput Intell Neurosci. 2016;2016:8402127. doi: 10.1155/2016/8402127. Epub 2015 Dec 27.

本文引用的文献

1
Representation learning: a review and new perspectives.表示学习:综述与新视角。
IEEE Trans Pattern Anal Mach Intell. 2013 Aug;35(8):1798-828. doi: 10.1109/TPAMI.2013.50.
2
Reducing the dimensionality of data with neural networks.使用神经网络降低数据维度。
Science. 2006 Jul 28;313(5786):504-7. doi: 10.1126/science.1127647.
3
A fast learning algorithm for deep belief nets.一种用于深度信念网络的快速学习算法。
Neural Comput. 2006 Jul;18(7):1527-54. doi: 10.1162/neco.2006.18.7.1527.
4
Training products of experts by minimizing contrastive divergence.通过最小化对比散度来训练专家的产品。
Neural Comput. 2002 Aug;14(8):1771-800. doi: 10.1162/089976602760128018.