• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

探讨改进 BERT 模型在生物医学关系抽取中的预训练和微调。

Investigation of improving the pre-training and fine-tuning of BERT model for biomedical relation extraction.

机构信息

Department of Computer and Information Science, Biomedical Text Mining Lab, University of Delaware, Newark, USA.

出版信息

BMC Bioinformatics. 2022 Apr 4;23(1):120. doi: 10.1186/s12859-022-04642-w.

DOI:10.1186/s12859-022-04642-w
PMID:35379166
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8978438/
Abstract

BACKGROUND

Recently, automatically extracting biomedical relations has been a significant subject in biomedical research due to the rapid growth of biomedical literature. Since the adaptation to the biomedical domain, the transformer-based BERT models have produced leading results on many biomedical natural language processing tasks. In this work, we will explore the approaches to improve the BERT model for relation extraction tasks in both the pre-training and fine-tuning stages of its applications. In the pre-training stage, we add another level of BERT adaptation on sub-domain data to bridge the gap between domain knowledge and task-specific knowledge. Also, we propose methods to incorporate the ignored knowledge in the last layer of BERT to improve its fine-tuning.

RESULTS

The experiment results demonstrate that our approaches for pre-training and fine-tuning can improve the BERT model performance. After combining the two proposed techniques, our approach outperforms the original BERT models with averaged F1 score improvement of 2.1% on relation extraction tasks. Moreover, our approach achieves state-of-the-art performance on three relation extraction benchmark datasets.

CONCLUSIONS

The extra pre-training step on sub-domain data can help the BERT model generalization on specific tasks, and our proposed fine-tuning mechanism could utilize the knowledge in the last layer of BERT to boost the model performance. Furthermore, the combination of these two approaches further improves the performance of BERT model on the relation extraction tasks.

摘要

背景

由于生物医学文献的快速增长,最近自动提取生物医学关系已成为生物医学研究中的一个重要课题。自从适应生物医学领域以来,基于转换器的 BERT 模型在许多生物医学自然语言处理任务上取得了领先的结果。在这项工作中,我们将探索改进 BERT 模型的方法,以提高其在应用的预训练和微调阶段的关系提取任务的性能。在预训练阶段,我们在子域数据上添加了另一层 BERT 适配,以弥合领域知识和特定于任务的知识之间的差距。此外,我们提出了在 BERT 的最后一层中纳入忽略知识的方法,以提高其微调效果。

结果

实验结果表明,我们的预训练和微调方法可以提高 BERT 模型的性能。在结合了这两种提出的技术后,我们的方法在关系提取任务上的平均 F1 得分提高了 2.1%,优于原始的 BERT 模型。此外,我们的方法在三个关系提取基准数据集上达到了最新的性能。

结论

在子域数据上进行额外的预训练步骤可以帮助 BERT 模型在特定任务上进行泛化,我们提出的微调机制可以利用 BERT 最后一层的知识来提高模型性能。此外,这两种方法的结合进一步提高了 BERT 模型在关系提取任务上的性能。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/03b1/8978438/4496b17f38f5/12859_2022_4642_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/03b1/8978438/23f7e9305c92/12859_2022_4642_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/03b1/8978438/93743f2eae17/12859_2022_4642_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/03b1/8978438/99c5a13a83a5/12859_2022_4642_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/03b1/8978438/dcd486d06738/12859_2022_4642_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/03b1/8978438/4496b17f38f5/12859_2022_4642_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/03b1/8978438/23f7e9305c92/12859_2022_4642_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/03b1/8978438/93743f2eae17/12859_2022_4642_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/03b1/8978438/99c5a13a83a5/12859_2022_4642_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/03b1/8978438/dcd486d06738/12859_2022_4642_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/03b1/8978438/4496b17f38f5/12859_2022_4642_Fig5_HTML.jpg

相似文献

1
Investigation of improving the pre-training and fine-tuning of BERT model for biomedical relation extraction.探讨改进 BERT 模型在生物医学关系抽取中的预训练和微调。
BMC Bioinformatics. 2022 Apr 4;23(1):120. doi: 10.1186/s12859-022-04642-w.
2
BioBERT and Similar Approaches for Relation Extraction.BioBERT 及其在关系抽取中的应用。
Methods Mol Biol. 2022;2496:221-235. doi: 10.1007/978-1-0716-2305-3_12.
3
BioBERT: a pre-trained biomedical language representation model for biomedical text mining.BioBERT:一种用于生物医学文本挖掘的预训练生物医学语言表示模型。
Bioinformatics. 2020 Feb 15;36(4):1234-1240. doi: 10.1093/bioinformatics/btz682.
4
BertSRC: transformer-based semantic relation classification.BertSRC:基于转换器的语义关系分类。
BMC Med Inform Decis Mak. 2022 Sep 6;22(1):234. doi: 10.1186/s12911-022-01977-5.
5
Extracting comprehensive clinical information for breast cancer using deep learning methods.利用深度学习方法提取乳腺癌全面临床信息。
Int J Med Inform. 2019 Dec;132:103985. doi: 10.1016/j.ijmedinf.2019.103985. Epub 2019 Oct 2.
6
CACER: Clinical concept Annotations for Cancer Events and Relations.CACER:癌症事件与关系的临床概念注释。
J Am Med Inform Assoc. 2024 Nov 1;31(11):2583-2594. doi: 10.1093/jamia/ocae231.
7
Benchmarking for biomedical natural language processing tasks with a domain specific ALBERT.基于领域特定的 ALBERT 进行生物医学自然语言处理任务的基准测试。
BMC Bioinformatics. 2022 Apr 21;23(1):144. doi: 10.1186/s12859-022-04688-w.
8
A general approach for improving deep learning-based medical relation extraction using a pre-trained model and fine-tuning.一种使用预训练模型和微调改进基于深度学习的医学关系抽取的通用方法。
Database (Oxford). 2019 Jan 1;2019. doi: 10.1093/database/baz116.
9
BERT-GT: cross-sentence n-ary relation extraction with BERT and Graph Transformer.BERT-GT:使用BERT和图变换器进行跨句子n元关系提取
Bioinformatics. 2021 Apr 5;36(24):5678-5685. doi: 10.1093/bioinformatics/btaa1087.
10
Biomedical relation extraction method based on ensemble learning and attention mechanism.基于集成学习和注意力机制的生物医学关系抽取方法。
BMC Bioinformatics. 2024 Oct 18;25(1):333. doi: 10.1186/s12859-024-05951-y.

引用本文的文献

1
Predicting 30-Day Postoperative Mortality and American Society of Anesthesiologists Physical Status Using Retrieval-Augmented Large Language Models: Development and Validation Study.使用检索增强大语言模型预测术后30天死亡率和美国麻醉医师协会身体状况:开发与验证研究
J Med Internet Res. 2025 Jun 3;27:e75052. doi: 10.2196/75052.
2
Iterative refinement and goal articulation to optimize large language models for clinical information extraction.迭代优化与目标阐述以优化用于临床信息提取的大语言模型
NPJ Digit Med. 2025 May 23;8(1):301. doi: 10.1038/s41746-025-01686-z.
3
GlycoSiteMiner: an ML/AI-assisted literature mining-based pipeline for extracting glycosylation sites from PubMed abstracts.

本文引用的文献

1
Fine-tuning large neural language models for biomedical natural language processing.针对生物医学自然语言处理对大型神经语言模型进行微调。
Patterns (N Y). 2023 Apr 14;4(4):100729. doi: 10.1016/j.patter.2023.100729.
2
ProteinBERT: a universal deep-learning model of protein sequence and function.蛋白质 BERT:一种通用的蛋白质序列和功能深度学习模型。
Bioinformatics. 2022 Apr 12;38(8):2102-2110. doi: 10.1093/bioinformatics/btac020.
3
BioBERT: a pre-trained biomedical language representation model for biomedical text mining.BioBERT:一种用于生物医学文本挖掘的预训练生物医学语言表示模型。
糖基位点挖掘工具(GlycoSiteMiner):一种基于机器学习/人工智能辅助文献挖掘的流程,用于从PubMed摘要中提取糖基化位点。
Glycobiology. 2025 Jun 2;35(7). doi: 10.1093/glycob/cwaf030.
4
Prospects and perils of ChatGPT in diabetes.ChatGPT在糖尿病领域的前景与风险
World J Diabetes. 2025 Mar 15;16(3):98408. doi: 10.4239/wjd.v16.i3.98408.
5
Prompts to Table: Specification and Iterative Refinement for Clinical Information Extraction with Large Language Models.表格提示:用于使用大语言模型进行临床信息提取的规范与迭代优化
medRxiv. 2025 Apr 1:2025.02.11.25322107. doi: 10.1101/2025.02.11.25322107.
6
Biomedical relation extraction method based on ensemble learning and attention mechanism.基于集成学习和注意力机制的生物医学关系抽取方法。
BMC Bioinformatics. 2024 Oct 18;25(1):333. doi: 10.1186/s12859-024-05951-y.
7
LLM-AIx: An open source pipeline for Information Extraction from unstructured medical text based on privacy preserving Large Language Models.LLM-AIx:一种基于隐私保护大语言模型从非结构化医学文本中提取信息的开源管道。
medRxiv. 2024 Sep 3:2024.09.02.24312917. doi: 10.1101/2024.09.02.24312917.
8
Automatic extraction of transcriptional regulatory interactions of bacteria from biomedical literature using a BERT-based approach.基于 BERT 的方法从生物医学文献中自动提取细菌的转录调控相互作用。
Database (Oxford). 2024 Aug 30;2024. doi: 10.1093/database/baae094.
9
A deep learning-driven discovery of berberine derivatives as novel antibacterial against multidrug-resistant Helicobacter pylori.基于深度学习的黄连素衍生物的发现:一种新型抗多重耐药幽门螺杆菌的抗菌药物。
Signal Transduct Target Ther. 2024 Jul 8;9(1):183. doi: 10.1038/s41392-024-01895-0.
10
Leveraging Unlabeled Clinical Data to Boost Performance of Risk Stratification Models for Suspected Acute Coronary Syndrome.利用未标记的临床数据提高疑似急性冠状动脉综合征风险分层模型的性能。
AMIA Annu Symp Proc. 2024 Jan 11;2023:744-753. eCollection 2023.
Bioinformatics. 2020 Feb 15;36(4):1234-1240. doi: 10.1093/bioinformatics/btz682.
4
Using distant supervision to augment manually annotated data for relation extraction.利用远程监督来扩充人工标注数据以进行关系抽取。
PLoS One. 2019 Jul 30;14(7):e0216913. doi: 10.1371/journal.pone.0216913. eCollection 2019.
5
Drug-drug interaction extraction from biomedical texts using long short-term memory network.基于长短时记忆网络的生物医学文献中药物-药物相互作用提取
J Biomed Inform. 2018 Oct;86:15-24. doi: 10.1016/j.jbi.2018.08.005. Epub 2018 Aug 21.
6
MIMIC-III, a freely accessible critical care database.MIMIC-III,一个免费获取的重症监护数据库。
Sci Data. 2016 May 24;3:160035. doi: 10.1038/sdata.2016.35.
7
The DDI corpus: an annotated corpus with pharmacological substances and drug-drug interactions.DDI 语料库:一个带有药理学物质和药物相互作用注释的语料库。
J Biomed Inform. 2013 Oct;46(5):914-20. doi: 10.1016/j.jbi.2013.07.011. Epub 2013 Jul 29.
8
PubTator: a web-based text mining tool for assisting biocuration.PubTator:一个用于辅助生物注释的基于网络的文本挖掘工具。
Nucleic Acids Res. 2013 Jul;41(Web Server issue):W518-22. doi: 10.1093/nar/gkt441. Epub 2013 May 22.
9
Overview of the protein-protein interaction annotation extraction task of BioCreative II.生物创意II蛋白质-蛋白质相互作用注释提取任务概述。
Genome Biol. 2008;9 Suppl 2(Suppl 2):S4. doi: 10.1186/gb-2008-9-s2-s4. Epub 2008 Sep 1.
10
Comparative experiments on learning information extractors for proteins and their interactions.蛋白质及其相互作用的学习信息提取器的比较实验。
Artif Intell Med. 2005 Feb;33(2):139-55. doi: 10.1016/j.artmed.2004.07.016.