• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

B-LBConA:一种基于 Bio-LinkBERT 和上下文感知机制的医学实体消歧模型。

B-LBConA: a medical entity disambiguation model based on Bio-LinkBERT and context-aware mechanism.

机构信息

Key Laboratory of Advanced Design and Intelligent Computing, Ministry of Education, Dalian University, 116622, Dalian, China.

School of Computer Science and Artificial Intelligence, Wuhan University of Technology, 430070, Wuhan, China.

出版信息

BMC Bioinformatics. 2023 Mar 16;24(1):97. doi: 10.1186/s12859-023-05209-z.

DOI:10.1186/s12859-023-05209-z
PMID:36927359
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10021986/
Abstract

BACKGROUND

The main task of medical entity disambiguation is to link mentions, such as diseases, drugs, or complications, to standard entities in the target knowledge base. To our knowledge, models based on Bidirectional Encoder Representations from Transformers (BERT) have achieved good results in this task. Unfortunately, these models only consider text in the current document, fail to capture dependencies with other documents, and lack sufficient mining of hidden information in contextual texts.

RESULTS

We propose B-LBConA, which is based on Bio-LinkBERT and context-aware mechanism. Specifically, B-LBConA first utilizes Bio-LinkBERT, which is capable of learning cross-document dependencies, to obtain embedding representations of mentions and candidate entities. Then, cross-attention is used to capture the interaction information of mention-to-entity and entity-to-mention. Finally, B-LBConA incorporates disambiguation clues about the relevance between the mention context and candidate entities via the context-aware mechanism.

CONCLUSIONS

Experiment results on three publicly available datasets, NCBI, ADR and ShARe/CLEF, show that B-LBConA achieves a signifcantly more accurate performance compared with existing models.

摘要

背景

医学实体消歧的主要任务是将提及(如疾病、药物或并发症)链接到目标知识库中的标准实体。据我们所知,基于变压器的双向编码器表示(BERT)的模型在这项任务中取得了很好的效果。不幸的是,这些模型仅考虑当前文档中的文本,无法捕获与其他文档的依赖关系,并且缺乏对上下文文本中隐藏信息的充分挖掘。

结果

我们提出了基于生物链接 BERT 和上下文感知机制的 B-LBConA。具体来说,B-LBConA 首先利用能够学习跨文档依赖关系的生物链接 BERT,获取提及和候选实体的嵌入表示。然后,使用交叉注意力来捕获提及到实体和实体到提及的交互信息。最后,B-LBConA 通过上下文感知机制结合提及上下文和候选实体之间相关性的消歧线索。

结论

在三个公开可用的数据集 NCBI、ADR 和 ShARe/CLEF 上的实验结果表明,B-LBConA 与现有模型相比,具有更高的准确性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9cee/10021986/c3b5eb81fd5c/12859_2023_5209_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9cee/10021986/d163c3cba9b0/12859_2023_5209_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9cee/10021986/b333e6625c9d/12859_2023_5209_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9cee/10021986/83da1509d411/12859_2023_5209_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9cee/10021986/c3b5eb81fd5c/12859_2023_5209_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9cee/10021986/d163c3cba9b0/12859_2023_5209_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9cee/10021986/b333e6625c9d/12859_2023_5209_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9cee/10021986/83da1509d411/12859_2023_5209_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9cee/10021986/c3b5eb81fd5c/12859_2023_5209_Fig4_HTML.jpg

相似文献

1
B-LBConA: a medical entity disambiguation model based on Bio-LinkBERT and context-aware mechanism.B-LBConA:一种基于 Bio-LinkBERT 和上下文感知机制的医学实体消歧模型。
BMC Bioinformatics. 2023 Mar 16;24(1):97. doi: 10.1186/s12859-023-05209-z.
2
Enhancing unsupervised medical entity linking with multi-instance learning.利用多实例学习增强无监督医学实体链接。
BMC Med Inform Decis Mak. 2021 Nov 16;21(Suppl 9):317. doi: 10.1186/s12911-021-01654-z.
3
Distinguished representation of identical mentions in bio-entity coreference resolution.生物实体共指消解中相同提及的出色表示。
BMC Med Inform Decis Mak. 2022 Apr 30;22(1):116. doi: 10.1186/s12911-022-01862-1.
4
A Knowledge Graph Entity Disambiguation Method Based on Entity-Relationship Embedding and Graph Structure Embedding.基于实体关系嵌入和图结构嵌入的知识图谱实体消歧方法。
Comput Intell Neurosci. 2021 Sep 23;2021:2878189. doi: 10.1155/2021/2878189. eCollection 2021.
5
LBERT: Lexically aware Transformer-based Bidirectional Encoder Representation model for learning universal bio-entity relations.LBERT:基于词汇感知的基于Transformer的双向编码器表示模型,用于学习通用生物实体关系。
Bioinformatics. 2021 Apr 20;37(3):404-412. doi: 10.1093/bioinformatics/btaa721.
6
Knowledge Graph Completion for the Chinese Text of Cultural Relics Based on Bidirectional Encoder Representations from Transformers with Entity-Type Information.基于带有实体类型信息的变换器双向编码器表征的文物中文文本知识图谱补全
Entropy (Basel). 2020 Oct 16;22(10):1168. doi: 10.3390/e22101168.
7
Knowledge based word-concept model estimation and refinement for biomedical text mining.用于生物医学文本挖掘的基于知识的词概念模型估计与优化。
J Biomed Inform. 2015 Feb;53:300-7. doi: 10.1016/j.jbi.2014.11.015. Epub 2014 Dec 12.
8
BELHD: improving biomedical entity linking with homonym disambiguation.BELHD:利用同形词消歧改进生物医学实体链接。
Bioinformatics. 2024 Aug 2;40(8). doi: 10.1093/bioinformatics/btae474.
9
Multi-Level Representation Learning for Chinese Medical Entity Recognition: Model Development and Validation.用于中文医学实体识别的多层次表示学习:模型开发与验证
JMIR Med Inform. 2020 May 4;8(5):e17637. doi: 10.2196/17637.
10
A metric learning-based method for biomedical entity linking.一种基于度量学习的生物医学实体链接方法。
Front Res Metr Anal. 2023 Dec 19;8:1247094. doi: 10.3389/frma.2023.1247094. eCollection 2023.

引用本文的文献

1
Medical named entity recognition based on domain knowledge and position encoding.基于领域知识和位置编码的医学命名实体识别
BMC Med Inform Decis Mak. 2025 Jul 1;25(1):235. doi: 10.1186/s12911-025-03037-0.

本文引用的文献

1
On the effectiveness of compact biomedical transformers.紧凑型生物医学变压器的有效性。
Bioinformatics. 2023 Mar 1;39(3). doi: 10.1093/bioinformatics/btad103.
2
Improving broad-coverage medical entity linking with semantic type prediction and large-scale datasets.利用语义类型预测和大规模数据集提高全面的医学实体链接。
J Biomed Inform. 2021 Sep;121:103880. doi: 10.1016/j.jbi.2021.103880. Epub 2021 Aug 12.
3
BERT-based Ranking for Biomedical Entity Normalization.基于BERT的生物医学实体规范化排序
AMIA Jt Summits Transl Sci Proc. 2020 May 30;2020:269-277. eCollection 2020.
4
BioBERT: a pre-trained biomedical language representation model for biomedical text mining.BioBERT:一种用于生物医学文本挖掘的预训练生物医学语言表示模型。
Bioinformatics. 2020 Feb 15;36(4):1234-1240. doi: 10.1093/bioinformatics/btz682.
5
CNN-based ranking for biomedical entity normalization.基于卷积神经网络的生物医学实体标准化排序
BMC Bioinformatics. 2017 Oct 3;18(Suppl 11):385. doi: 10.1186/s12859-017-1805-7.
6
TaggerOne: joint named entity recognition and normalization with semi-Markov Models.TaggerOne:使用半马尔可夫模型进行联合命名实体识别与归一化
Bioinformatics. 2016 Sep 15;32(18):2839-46. doi: 10.1093/bioinformatics/btw343. Epub 2016 Jun 9.
7
DNorm: disease name normalization with pairwise learning to rank.DNorm:基于对分学习排序的疾病名称标准化。
Bioinformatics. 2013 Nov 15;29(22):2909-17. doi: 10.1093/bioinformatics/btt474. Epub 2013 Aug 21.
8
Using rule-based natural language processing to improve disease normalization in biomedical text.基于规则的自然语言处理在生物医学文本疾病标准化中的应用。
J Am Med Inform Assoc. 2013 Sep-Oct;20(5):876-81. doi: 10.1136/amiajnl-2012-001173. Epub 2012 Oct 6.
9
The Unified Medical Language System (UMLS): integrating biomedical terminology.统一医学语言系统(UMLS):整合生物医学术语。
Nucleic Acids Res. 2004 Jan 1;32(Database issue):D267-70. doi: 10.1093/nar/gkh061.