• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过正则化注入本体知识来增强问答

Enhancing Question Answering by Injecting Ontological Knowledge through Regularization.

作者信息

Goodwin Travis R, Demner-Fushman Dina

机构信息

U.S. National Library of Medicine, National Institutes of Health.

出版信息

Proc Conf Empir Methods Nat Lang Process. 2020 Nov;2020:56-63. doi: 10.18653/v1/2020.deelio-1.7.

DOI:10.18653/v1/2020.deelio-1.7
PMID:33364628
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7757122/
Abstract

Deep neural networks have demonstrated high performance on many natural language processing (NLP) tasks that can be answered directly from text, and have struggled to solve NLP tasks requiring external (e.g., world) knowledge. In this paper, we present OSCR (Ontology-based Semantic Composition Regularization), a method for injecting task-agnostic knowledge from an Ontology or knowledge graph into a neural network during pre-training. We evaluated the performance of BERT pre-trained on Wikipedia with and without OSCR by measuring the performance when fine-tuning on two question answering tasks involving world knowledge and causal reasoning and one requiring domain (healthcare) knowledge and obtained 33.3 %, 18.6 %, and 4 % improved accuracy compared to pre-training BERT without OSCR.

摘要

深度神经网络在许多可直接从文本回答的自然语言处理(NLP)任务上已展现出高性能,但在解决需要外部(如世界)知识的NLP任务时却面临困难。在本文中,我们提出了基于本体的语义组合正则化(OSCR)方法,这是一种在预训练期间将来自本体或知识图谱的与任务无关的知识注入神经网络的方法。我们通过在两个涉及世界知识和因果推理以及一个需要领域(医疗保健)知识的问答任务上进行微调时测量性能,评估了在有和没有OSCR的情况下在维基百科上预训练的BERT的性能,与没有OSCR预训练的BERT相比,准确率分别提高了33.3%、18.6%和4%。

相似文献

1
Enhancing Question Answering by Injecting Ontological Knowledge through Regularization.通过正则化注入本体知识来增强问答
Proc Conf Empir Methods Nat Lang Process. 2020 Nov;2020:56-63. doi: 10.18653/v1/2020.deelio-1.7.
2
Incorporating Domain Knowledge Into Language Models by Using Graph Convolutional Networks for Assessing Semantic Textual Similarity: Model Development and Performance Comparison.通过使用图卷积网络将领域知识融入语言模型以评估语义文本相似度:模型开发与性能比较
JMIR Med Inform. 2021 Nov 26;9(11):e23101. doi: 10.2196/23101.
3
Infusing Multi-Hop Medical Knowledge Into Smaller Language Models for Biomedical Question Answering.将多跳医学知识融入小型语言模型以进行生物医学问答
IEEE J Biomed Health Inform. 2025 Mar 19;PP. doi: 10.1109/JBHI.2025.3547444.
4
A Question-and-Answer System to Extract Data From Free-Text Oncological Pathology Reports (CancerBERT Network): Development Study.从自由文本肿瘤病理学报告(CancerBERT 网络)中提取数据的问答系统:开发研究。
J Med Internet Res. 2022 Mar 23;24(3):e27210. doi: 10.2196/27210.
5
When BERT meets Bilbo: a learning curve analysis of pretrained language model on disease classification.当 BERT 遇见比尔博:预训练语言模型在疾病分类上的学习曲线分析。
BMC Med Inform Decis Mak. 2022 Apr 5;21(Suppl 9):377. doi: 10.1186/s12911-022-01829-2.
6
Few-Shot Learning for Clinical Natural Language Processing Using Siamese Neural Networks: Algorithm Development and Validation Study.使用暹罗神经网络的临床自然语言处理少样本学习:算法开发与验证研究
JMIR AI. 2023 May 4;2:e44293. doi: 10.2196/44293.
7
BioBERT: a pre-trained biomedical language representation model for biomedical text mining.BioBERT:一种用于生物医学文本挖掘的预训练生物医学语言表示模型。
Bioinformatics. 2020 Feb 15;36(4):1234-1240. doi: 10.1093/bioinformatics/btz682.
8
COVID-Twitter-BERT: A natural language processing model to analyse COVID-19 content on Twitter.COVID-Twitter-BERT:一种用于分析推特上新冠疫情相关内容的自然语言处理模型。
Front Artif Intell. 2023 Mar 14;6:1023281. doi: 10.3389/frai.2023.1023281. eCollection 2023.
9
External features enriched model for biomedical question answering.生物医学问答的外部特征丰富模型。
BMC Bioinformatics. 2021 May 26;22(1):272. doi: 10.1186/s12859-021-04176-7.
10
Investigation of improving the pre-training and fine-tuning of BERT model for biomedical relation extraction.探讨改进 BERT 模型在生物医学关系抽取中的预训练和微调。
BMC Bioinformatics. 2022 Apr 4;23(1):120. doi: 10.1186/s12859-022-04642-w.

引用本文的文献

1
Development and application of Chinese medical ontology for diabetes mellitus.中文医学本体在糖尿病领域的开发与应用。
BMC Med Inform Decis Mak. 2024 Jan 19;24(1):18. doi: 10.1186/s12911-023-02405-y.

本文引用的文献

1
Embedding Open-domain Common-sense Knowledge from Text.从文本中嵌入开放域常识知识。
LREC Int Conf Lang Resour Eval. 2016 May;2016:4621-4628.
2
Recognizing Question Entailment for Medical Question Answering.识别医学问答中的问题蕴含关系。
AMIA Annu Symp Proc. 2017 Feb 10;2016:310-318. eCollection 2016.