Wang Yuanlong, Yin Changchang, Zhang Ping
Department of Computer Science and Engineering, The Ohio State University, Columbus, OH 43210, USA.
Department of Biomedical Informatics, The Ohio State University, Columbus, OH 43210, USA.
Heliyon. 2024 Feb 28;10(5):e26772. doi: 10.1016/j.heliyon.2024.e26772. eCollection 2024 Mar 15.
The broad adoption of electronic health record (EHR) systems brings us a tremendous amount of clinical data and thus provides opportunities to conduct data-based healthcare research to solve various clinical problems in the medical domain. Machine learning and deep learning methods are widely used in the medical informatics and healthcare domain due to their power to mine insights from raw data. When adapting deep learning models for EHR data, it is essential to consider its heterogeneous nature: EHR contains patient records from various sources including medical tests (e.g. blood test, microbiology test), medical imaging, diagnosis, medications, procedures, clinical notes, etc. Those modalities together provide a holistic view of patient health status and complement each other. Therefore, combining data from multiple modalities that are intrinsically different is challenging but intuitively promising in deep learning for EHR. To assess the expectations of multimodal data, we introduce a comprehensive fusion framework designed to integrate temporal variables, medical images, and clinical notes in EHR for enhanced performance in clinical risk prediction. Early, joint, and late fusion strategies are employed to combine data from various modalities effectively. We test the model with three predictive tasks: in-hospital mortality, long length of stay, and 30-day readmission. Experimental results show that multimodal models outperform uni-modal models in the tasks involved. Additionally, by training models with different input modality combinations, we calculate the Shapley value for each modality to quantify their contribution to multimodal performance. It is shown that temporal variables tend to be more helpful than CXR images and clinical notes in the three explored predictive tasks.
电子健康记录(EHR)系统的广泛应用为我们带来了大量临床数据,从而为开展基于数据的医疗保健研究以解决医学领域的各种临床问题提供了机会。机器学习和深度学习方法因其能够从原始数据中挖掘见解的能力而在医学信息学和医疗保健领域得到广泛应用。在将深度学习模型应用于EHR数据时,必须考虑其异构性质:EHR包含来自各种来源的患者记录,包括医学检查(如血液检查、微生物学检查)、医学影像、诊断、药物、手术、临床记录等。这些模态共同提供了患者健康状况的整体视图,并且相互补充。因此,在EHR的深度学习中,将本质上不同的多模态数据进行组合具有挑战性,但直观上很有前景。为了评估对多模态数据的期望,我们引入了一个综合融合框架,旨在整合EHR中的时间变量、医学影像和临床记录,以提高临床风险预测的性能。采用早期、联合和晚期融合策略来有效组合来自各种模态的数据。我们用三个预测任务对模型进行测试:院内死亡率、长期住院和30天再入院。实验结果表明,多模态模型在涉及的任务中优于单模态模型。此外,通过使用不同的输入模态组合训练模型,我们计算每个模态的沙普利值以量化它们对多模态性能的贡献。结果表明,在三个探索的预测任务中,时间变量往往比胸部X光图像和临床记录更有帮助。