• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

使用BEHRT进行医学概念嵌入的联邦学习。

Federated learning of medical concepts embedding using BEHRT.

作者信息

Ben Shoham Ofir, Rappoport Nadav

机构信息

Department of Software and Information Systems Engineering, Ben-Gurion University of the Negev, Be'er Sheva, Israel.

出版信息

JAMIA Open. 2024 Oct 23;7(4):ooae110. doi: 10.1093/jamiaopen/ooae110. eCollection 2024 Dec.

DOI:10.1093/jamiaopen/ooae110
PMID:39445033
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11498200/
Abstract

OBJECTIVES

Electronic health record data is often considered sensitive medical information. Therefore, the EHR data from different medical centers often cannot be shared, making it difficult to create prediction models using multicenter EHR data, which is essential for such models' robustness and generalizability. Federated learning (FL) is an algorithmic approach that allows learning a shared model using data in multiple locations without the need to store all data in a single central place. Our study aims to evaluate an FL approach using the BEHRT model for predictive tasks on EHR data, focusing on next visit prediction.

MATERIALS AND METHODS

We propose an FL approach for learning medical concepts embedding. This pretrained model can be used for fine-tuning for specific downstream tasks. Our approach is based on an embedding model like BEHRT, a deep neural sequence transduction model for EHR. We train using FL, both the masked language modeling (MLM) and the next visit downstream model.

RESULTS

We demonstrate our approach on the MIMIC-IV dataset. We compare the performance of a model trained with FL to one trained on centralized data, observing a difference in average precision ranging from 0% to 3% (absolute), depending on the length of the patients' visit history. Moreover, our approach improves average precision by 4%-10% (absolute) compared to local models. In addition, we show the importance of the usage of pretrained MLM for the next visit diagnoses prediction task.

DISCUSSION AND CONCLUSION

We find that our FL approach reaches very close to the performance of a centralized model, and it outperforms local models in terms of average precision. We also show that pretrained MLM improves the model's average precision performance in the next visit diagnoses prediction task, compared to an MLM without pretraining.

摘要

目标

电子健康记录数据通常被视为敏感的医疗信息。因此,来自不同医疗中心的电子健康记录数据往往无法共享,这使得利用多中心电子健康记录数据创建预测模型变得困难,而这对于此类模型的稳健性和通用性至关重要。联邦学习(FL)是一种算法方法,它允许使用多个位置的数据学习共享模型,而无需将所有数据存储在单个中心位置。我们的研究旨在评估一种使用BEHRT模型的联邦学习方法,用于电子健康记录数据的预测任务,重点是下次就诊预测。

材料与方法

我们提出了一种用于学习医学概念嵌入的联邦学习方法。这个预训练模型可用于针对特定下游任务进行微调。我们的方法基于一种类似BEHRT的嵌入模型,这是一种用于电子健康记录的深度神经序列转导模型。我们使用联邦学习来训练掩码语言模型(MLM)和下次就诊下游模型。

结果

我们在MIMIC-IV数据集上展示了我们的方法。我们将使用联邦学习训练的模型与在集中式数据上训练的模型的性能进行比较,观察到平均精度的差异在0%至3%(绝对值)之间,具体取决于患者就诊历史的长度。此外,与本地模型相比,我们的方法将平均精度提高了4% - 10%(绝对值)。此外,我们展示了预训练的MLM在下次就诊诊断预测任务中的重要性。

讨论与结论

我们发现我们的联邦学习方法的性能非常接近集中式模型,并且在平均精度方面优于本地模型。我们还表明,与未进行预训练的MLM相比,预训练的MLM在下次就诊诊断预测任务中提高了模型的平均精度性能。

相似文献

1
Federated learning of medical concepts embedding using BEHRT.使用BEHRT进行医学概念嵌入的联邦学习。
JAMIA Open. 2024 Oct 23;7(4):ooae110. doi: 10.1093/jamiaopen/ooae110. eCollection 2024 Dec.
2
Disease Concept-Embedding Based on the Self-Supervised Method for Medical Information Extraction from Electronic Health Records and Disease Retrieval: Algorithm Development and Validation Study.基于自监督方法的疾病概念嵌入在电子健康记录中的医学信息提取和疾病检索:算法开发和验证研究。
J Med Internet Res. 2021 Jan 27;23(1):e25113. doi: 10.2196/25113.
3
Secure Extraction of Personal Information from EHR by Federated Machine Learning.联邦机器学习从电子健康记录中安全提取个人信息。
Stud Health Technol Inform. 2024 Aug 22;316:611-615. doi: 10.3233/SHTI240488.
4
Predicting treatment response in multicenter non-small cell lung cancer patients based on federated learning.基于联邦学习预测多中心非小细胞肺癌患者的治疗反应。
BMC Cancer. 2024 Jun 5;24(1):688. doi: 10.1186/s12885-024-12456-7.
5
Training a Deep Contextualized Language Model for International Classification of Diseases, 10th Revision Classification via Federated Learning: Model Development and Validation Study.通过联邦学习训练用于国际疾病分类第10次修订版分类的深度情境化语言模型:模型开发与验证研究
JMIR Med Inform. 2022 Nov 10;10(11):e41342. doi: 10.2196/41342.
6
Med-BERT: pretrained contextualized embeddings on large-scale structured electronic health records for disease prediction.医学BERT:基于大规模结构化电子健康记录进行疾病预测的预训练上下文嵌入模型
NPJ Digit Med. 2021 May 20;4(1):86. doi: 10.1038/s41746-021-00455-y.
7
EHR-BERT: A BERT-based model for effective anomaly detection in electronic health records.EHR-BERT:一种基于 BERT 的电子健康记录中有效异常检测模型。
J Biomed Inform. 2024 Feb;150:104605. doi: 10.1016/j.jbi.2024.104605. Epub 2024 Feb 6.
8
BEHRT: Transformer for Electronic Health Records.BEHRT:电子健康记录的转换器。
Sci Rep. 2020 Apr 28;10(1):7155. doi: 10.1038/s41598-020-62922-y.
9
Decentralized collaborative multi-institutional PET attenuation and scatter correction using federated deep learning.利用联邦深度学习进行去中心化协作的多机构 PET 衰减和散射校正。
Eur J Nucl Med Mol Imaging. 2023 Mar;50(4):1034-1050. doi: 10.1007/s00259-022-06053-8. Epub 2022 Dec 12.
10
Performance of federated learning-based models in the Dutch TAVI population was comparable to central strategies and outperformed local strategies.基于联邦学习的模型在荷兰经导管主动脉瓣植入术人群中的表现与集中式策略相当,且优于局部策略。
Front Cardiovasc Med. 2024 Jul 5;11:1399138. doi: 10.3389/fcvm.2024.1399138. eCollection 2024.

引用本文的文献

1
CPLLM: Clinical prediction with large language models.CPLLM:基于大语言模型的临床预测
PLOS Digit Health. 2024 Dec 6;3(12):e0000680. doi: 10.1371/journal.pdig.0000680. eCollection 2024 Dec.

本文引用的文献

1
FedEHR: A Federated Learning Approach towards the Prediction of Heart Diseases in IoT-Based Electronic Health Records.联邦电子健康记录(FedEHR):一种基于物联网的电子健康记录中预测心脏病的联邦学习方法。
Diagnostics (Basel). 2023 Oct 10;13(20):3166. doi: 10.3390/diagnostics13203166.
2
Author Correction: MIMIC-IV, a freely accessible electronic health record dataset.作者更正:MIMIC-IV,一个可免费获取的电子健康记录数据集。
Sci Data. 2023 Apr 18;10(1):219. doi: 10.1038/s41597-023-02136-9.
3
Federated learning enables big data for rare cancer boundary detection.
联邦学习为罕见癌症边界检测提供大数据支持。
Nat Commun. 2022 Dec 5;13(1):7346. doi: 10.1038/s41467-022-33407-5.
4
Med-BERT: pretrained contextualized embeddings on large-scale structured electronic health records for disease prediction.医学BERT:基于大规模结构化电子健康记录进行疾病预测的预训练上下文嵌入模型
NPJ Digit Med. 2021 May 20;4(1):86. doi: 10.1038/s41746-021-00455-y.
5
Bidirectional Representation Learning From Transformers Using Multimodal Electronic Health Record Data to Predict Depression.利用多模态电子健康记录数据从转换器中进行双向表示学习以预测抑郁。
IEEE J Biomed Health Inform. 2021 Aug;25(8):3121-3129. doi: 10.1109/JBHI.2021.3063721. Epub 2021 Aug 5.
6
Deep representation learning of patient data from Electronic Health Records (EHR): A systematic review.电子健康记录(EHR)中患者数据的深度表征学习:一项系统综述。
J Biomed Inform. 2021 Mar;115:103671. doi: 10.1016/j.jbi.2020.103671. Epub 2020 Dec 31.
7
Federated Learning for Healthcare Informatics.医疗信息学中的联邦学习
J Healthc Inform Res. 2021;5(1):1-19. doi: 10.1007/s41666-020-00082-4. Epub 2020 Nov 12.
8
BEHRT: Transformer for Electronic Health Records.BEHRT:电子健康记录的转换器。
Sci Rep. 2020 Apr 28;10(1):7155. doi: 10.1038/s41598-020-62922-y.
9
Electronic Health Records: Then, Now, and in the Future.电子健康记录:过去、现在与未来。
Yearb Med Inform. 2016 May 20;Suppl 1(Suppl 1):S48-61. doi: 10.15265/IYS-2016-s006.
10
Deep Patient: An Unsupervised Representation to Predict the Future of Patients from the Electronic Health Records.深度患者:一种从电子健康记录中预测患者未来的无监督表示。
Sci Rep. 2016 May 17;6:26094. doi: 10.1038/srep26094.