• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于医疗保健应用的高效机器阅读理解:上下文提取方法的算法开发与验证

Efficient Machine Reading Comprehension for Health Care Applications: Algorithm Development and Validation of a Context Extraction Approach.

作者信息

Nguyen Duy-Anh, Li Minyi, Lambert Gavin, Kowalczyk Ryszard, McDonald Rachael, Vo Quoc Bao

机构信息

School of Software and Electrical Engineering, Swinburne University of Technology, Hawthorn, Australia.

School of Computing Technologies, RMIT, Melbourne, Australia.

出版信息

JMIR Form Res. 2024 Mar 25;8:e52482. doi: 10.2196/52482.

DOI:10.2196/52482
PMID:38526545
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11002730/
Abstract

BACKGROUND

Extractive methods for machine reading comprehension (MRC) tasks have achieved comparable or better accuracy than human performance on benchmark data sets. However, such models are not as successful when adapted to complex domains such as health care. One of the main reasons is that the context that the MRC model needs to process when operating in a complex domain can be much larger compared with an average open-domain context. This causes the MRC model to make less accurate and slower predictions. A potential solution to this problem is to reduce the input context of the MRC model by extracting only the necessary parts from the original context.

OBJECTIVE

This study aims to develop a method for extracting useful contexts from long articles as an additional component to the question answering task, enabling the MRC model to work more efficiently and accurately.

METHODS

Existing approaches to context extraction in MRC are based on sentence selection strategies, in which the models are trained to find the sentences containing the answer. We found that using only the sentences containing the answer was insufficient for the MRC model to predict correctly. We conducted a series of empirical studies and observed a strong relationship between the usefulness of the context and the confidence score output of the MRC model. Our investigation showed that a precise input context can boost the prediction correctness of the MRC and greatly reduce inference time. We proposed a method to estimate the utility of each sentence in a context in answering the question and then extract a new, shorter context according to these estimations. We generated a data set to train 2 models for estimating sentence utility, based on which we selected more precise contexts that improved the MRC model's performance.

RESULTS

We demonstrated our approach on the Question Answering Data Set for COVID-19 and Biomedical Semantic Indexing and Question Answering data sets and showed that the approach benefits the downstream MRC model. First, the method substantially reduced the inference time of the entire question answering system by 6 to 7 times. Second, our approach helped the MRC model predict the answer more correctly compared with using the original context (F-score increased from 0.724 to 0.744 for the Question Answering Data Set for COVID-19 and from 0.651 to 0.704 for the Biomedical Semantic Indexing and Question Answering). We also found a potential problem where extractive transformer MRC models predict poorly despite being given a more precise context in some cases.

CONCLUSIONS

The proposed context extraction method allows the MRC model to achieve improved prediction correctness and a significantly reduced MRC inference time. This approach works technically with any MRC model and has potential in tasks involving processing long texts.

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2401/11002730/199f2b2cd718/formative_v8i1e52482_fig3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2401/11002730/3d958ffc34f1/formative_v8i1e52482_fig1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2401/11002730/a676933e14ee/formative_v8i1e52482_fig2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2401/11002730/199f2b2cd718/formative_v8i1e52482_fig3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2401/11002730/3d958ffc34f1/formative_v8i1e52482_fig1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2401/11002730/a676933e14ee/formative_v8i1e52482_fig2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2401/11002730/199f2b2cd718/formative_v8i1e52482_fig3.jpg
摘要

背景

用于机器阅读理解(MRC)任务的提取方法在基准数据集上已取得与人类表现相当或更好的准确率。然而,当应用于医疗保健等复杂领域时,此类模型并不那么成功。主要原因之一是,与一般开放域上下文相比,MRC模型在复杂领域运行时需要处理的上下文可能要大得多。这导致MRC模型做出的预测准确性降低且速度变慢。解决此问题的一个潜在方法是通过仅从原始上下文中提取必要部分来减少MRC模型的输入上下文。

目的

本研究旨在开发一种从长篇文章中提取有用上下文的方法,作为问答任务的一个附加组件,使MRC模型能够更高效、准确地工作。

方法

MRC中现有的上下文提取方法基于句子选择策略,即训练模型找到包含答案的句子。我们发现仅使用包含答案的句子不足以让MRC模型正确预测。我们进行了一系列实证研究,并观察到上下文的有用性与MRC模型输出的置信度得分之间存在密切关系。我们的调查表明,精确的输入上下文可以提高MRC的预测正确性并大大减少推理时间。我们提出了一种方法来估计上下文中每个句子在回答问题时的效用,然后根据这些估计提取一个新的、更短的上下文。我们生成了一个数据集来训练2个用于估计句子效用的模型,并在此基础上选择了更精确的上下文,从而提高了MRC模型的性能。

结果

我们在新冠肺炎问答数据集以及生物医学语义索引与问答数据集上展示了我们的方法,并表明该方法对下游MRC模型有益。首先,该方法将整个问答系统的推理时间大幅减少了6至7倍。其次,与使用原始上下文相比,我们的方法帮助MRC模型更正确地预测答案(新冠肺炎问答数据集的F值从0.724提高到0.744,生物医学语义索引与问答数据集的F值从0.651提高到0.704)。我们还发现了一个潜在问题,即在某些情况下,尽管给定了更精确的上下文,但抽取式变压器MRC模型的预测效果仍然不佳。

结论

所提出的上下文提取方法使MRC模型能够提高预测正确性,并显著减少MRC推理时间。该方法在技术上适用于任何MRC模型,并且在涉及处理长文本的任务中具有潜力。

相似文献

1
Efficient Machine Reading Comprehension for Health Care Applications: Algorithm Development and Validation of a Context Extraction Approach.用于医疗保健应用的高效机器阅读理解:上下文提取方法的算法开发与验证
JMIR Form Res. 2024 Mar 25;8:e52482. doi: 10.2196/52482.
2
On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformers.基于生成式预训练变换器解决基于机器阅读理解的问答中的文本歧义与语义模糊问题。
PeerJ Comput Sci. 2023 Jul 24;9:e1422. doi: 10.7717/peerj-cs.1422. eCollection 2023.
3
HCT: Chinese Medical Machine Reading Comprehension Question-Answering via Hierarchically Collaborative Transformer.HCT:基于层次协作 Transformer 的中文医学机器阅读理解问答
IEEE J Biomed Health Inform. 2024 May;28(5):3055-3066. doi: 10.1109/JBHI.2024.3368288. Epub 2024 May 6.
4
Folic acid supplementation and malaria susceptibility and severity among people taking antifolate antimalarial drugs in endemic areas.在流行地区,服用抗叶酸抗疟药物的人群中,叶酸补充剂与疟疾易感性和严重程度的关系。
Cochrane Database Syst Rev. 2022 Feb 1;2(2022):CD014217. doi: 10.1002/14651858.CD014217.
5
Clinical concept and relation extraction using prompt-based machine reading comprehension.基于提示的机器阅读理解的临床概念和关系抽取。
J Am Med Inform Assoc. 2023 Aug 18;30(9):1486-1493. doi: 10.1093/jamia/ocad107.
6
MGRC: An End-to-End Multigranularity Reading Comprehension Model for Question Answering.MGRC:用于问答的端到端多粒度阅读理解模型
IEEE Trans Neural Netw Learn Syst. 2023 May;34(5):2594-2605. doi: 10.1109/TNNLS.2021.3107029. Epub 2023 May 2.
7
: a novel resource for question answering on scholarly articles.一种用于学术文章问答的新型资源。
Int J Digit Libr. 2022;23(3):289-301. doi: 10.1007/s00799-022-00329-y. Epub 2022 Jul 20.
8
A Pilot Study of Biomedical Text Comprehension using an Attention-Based Deep Neural Reader: Design and Experimental Analysis.一项使用基于注意力的深度神经阅读器进行生物医学文本理解的初步研究:设计与实验分析。
JMIR Med Inform. 2018 Jan 5;6(1):e2. doi: 10.2196/medinform.8751.
9
Deep learning-based approach for Arabic open domain question answering.基于深度学习的阿拉伯语开放域问答方法。
PeerJ Comput Sci. 2022 May 4;8:e952. doi: 10.7717/peerj-cs.952. eCollection 2022.
10
Predicting Semantic Similarity Between Clinical Sentence Pairs Using Transformer Models: Evaluation and Representational Analysis.使用Transformer模型预测临床句子对之间的语义相似性:评估与表征分析
JMIR Med Inform. 2021 May 26;9(5):e23099. doi: 10.2196/23099.

本文引用的文献

1
BioBERT: a pre-trained biomedical language representation model for biomedical text mining.BioBERT:一种用于生物医学文本挖掘的预训练生物医学语言表示模型。
Bioinformatics. 2020 Feb 15;36(4):1234-1240. doi: 10.1093/bioinformatics/btz682.