Song Feier, Liu Huazhang, Ma Huan, Chen Xuanhui, Wang Shouhong, Qin Tiehe, Liang Huiying, Huang Daozheng
Department of Emergency and Intensive Care Unit, Guangdong Provincial People's Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, 510080, China.
Medical Big Data Center, Guangdong Provincial People's Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Number 106, Zhongshaner Road, Guangzhou, 510080, China, 86 13416404410.
JMIR Form Res. 2025 Sep 8;9:e72482. doi: 10.2196/72482.
Point-of-care ultrasonography has become a valuable tool for assessing diaphragmatic function in critically ill patients receiving invasive mechanical ventilation. However, conventional diaphragm ultrasound assessment remains highly operator-dependent and subjective. Previous research introduced automatic measurement of diaphragmatic excursion and velocity using 2D speckle-tracking technology.
This study aimed to develop an artificial intelligence-multimodal learning framework to improve the prediction of weaning failure and guide individualized weaning strategies.
This prospective study enrolled critically ill patients older than 18 years who received mechanical ventilation for more than 48 hours and were eligible for a spontaneous breathing trial in 2 intensive care units in Guangzhou, China. Before the spontaneous breathing trial, diaphragm ultrasound videos were collected using a standardized protocol, and automatic measurements of excursion and velocity were obtained. A total of 88 patients were included, with 50 successfully weaned and 38 experiencing weaning failure. Each patient record included 27 clinical and 6 diaphragmatic indicators, selected based on previous literature and phenotyping studies. Clinical variables were preprocessed using OneHotEncoder, normalization, and scaling. Ultrasound videos were interpolated to a uniform resolution of 224×224×96. Artificial intelligence-multimodal learning based on clinical characteristics, laboratory parameters, and diaphragm ultrasonic videos was established. Four experiments were conducted in an ablation setting to evaluate model performance using different combinations of input data: (1) diaphragmatic excursion only, (2) clinical and diaphragmatic indicators, (3) ultrasound videos only, and (4) all modalities combined (multimodal). Metrics for evaluation included classification accuracy, area under the receiver operating characteristic curve (AUC), average precision in the precision-recall curve, and calibration curve. Variable importance was assessed using SHAP (Shapley Additive Explanation) to interpret feature contributions and understand model predictions.
The multimodal co-learning model outperformed all single-modal approaches. The accuracy improved when predicted through diaphragm ultrasound video data using Video Vision Transformer (accuracy=0.8095, AUC=0.852), clinical or ultrasound indicators (accuracy=0.7381, AUC=0.746), and the multimodal co-learning (accuracy=0.8331, AUC=0.894). The proposed co-learning model achieved the highest score (average precision=0.91) among the 4 experiments. Furthermore, calibration curve analysis demonstrated that the proposed colearning model was well calibrated, as the curve was closest to the perfectly calibrated line.
Combining ultrasound and clinical data for colearning improved the accuracy of the weaning outcome prediction. Multimodal learning based on automatic measurement of point-of-care ultrasonography and automated collection of objective clinical indicators greatly enhanced the practical operability and user-friendliness of the system. The proposed model offered promising potential for widespread clinical application in intensive care settings.
床旁超声已成为评估接受有创机械通气的危重症患者膈肌功能的重要工具。然而,传统的膈肌超声评估仍高度依赖操作者且主观性强。先前的研究引入了使用二维斑点追踪技术自动测量膈肌移动度和速度。
本研究旨在开发一种人工智能多模态学习框架,以改善对撤机失败的预测并指导个体化撤机策略。
这项前瞻性研究纳入了年龄大于18岁、接受机械通气超过48小时且符合在中国广州两家重症监护病房进行自主呼吸试验条件的危重症患者。在自主呼吸试验前,使用标准化方案收集膈肌超声视频,并获得移动度和速度的自动测量值。共纳入88例患者,其中50例成功撤机,38例撤机失败。每个患者记录包括基于先前文献和表型研究选择的27项临床指标和6项膈肌指标。临床变量使用独热编码器(OneHotEncoder)、归一化和缩放进行预处理。超声视频被插值到224×224×96的统一分辨率。建立了基于临床特征、实验室参数和膈肌超声视频的人工智能多模态学习模型。在消融设置下进行了四项实验,以使用不同的输入数据组合评估模型性能:(1)仅膈肌移动度,(2)临床和膈肌指标,(3)仅超声视频,(4)所有模态组合(多模态)。评估指标包括分类准确率、受试者操作特征曲线下面积(AUC)、精确召回曲线中的平均精度和校准曲线。使用SHAP(Shapley值加法解释)评估变量重要性,以解释特征贡献并理解模型预测。
多模态协同学习模型优于所有单模态方法。通过视频视觉变换器使用膈肌超声视频数据进行预测时准确率提高(准确率=0.8095,AUC=0.852),使用临床或超声指标时准确率为(准确率=0.7381,AUC=0.746),而多模态协同学习时准确率为(准确率=0.8331,AUC=0.894)。在四项实验中,所提出的协同学习模型获得了最高分(平均精度=0.91)。此外,校准曲线分析表明所提出的协同学习模型校准良好,因为该曲线最接近完美校准线。
将超声和临床数据结合进行协同学习提高了撤机结果预测的准确性。基于床旁超声自动测量和客观临床指标自动收集的多模态学习极大地提高了系统的实际可操作性和用户友好性。所提出的模型在重症监护环境中具有广泛临床应用的潜力。