Gao Riqiang, Tang Yucheng, Xu Kaiwen, Kammer Michael N, Antic Sanja L, Deppen Steve, Sandler Kim L, Massion Pierre P, Huo Yuankai, Landman Bennett A
Electrical Engineering and Computer Science, Vanderbilt University, Nashville, TN, USA 37235.
Vanderbilt University Medical Center, Nashville, TN, USA 37235.
Proc SPIE Int Soc Opt Eng. 2021 Feb;11596. doi: 10.1117/12.2580730. Epub 2021 Feb 15.
Clinical data elements (CDEs) (e.g., age, smoking history), blood markers and chest computed tomography (CT) structural features have been regarded as effective means for assessing lung cancer risk. These independent variables can provide complementary information and we hypothesize that combining them will improve the prediction accuracy. In practice, not all patients have all these variables available. In this paper, we propose a new network design, termed as multi-path multi-modal missing network (M3Net), to integrate the multi-modal data (i.e., CDEs, biomarker and CT image) considering missing modality with multiple paths neural network. Each path learns discriminative features of one modality, and different modalities are fused in a second stage for an integrated prediction. The network can be trained end-to-end with both medical image features and CDEs/biomarkers, or make a prediction with single modality. We evaluate M3Net with datasets including three sites from the Consortium for Molecular and Cellular Characterization of Screen-Detected Lesions (MCL) project. Our method is cross validated within a cohort of 1291 subjects (383 subjects with complete CDEs/biomarkers and CT images), and externally validated with a cohort of 99 subjects (99 with complete CDEs/biomarkers and CT images). Both cross-validation and external-validation results show that combining multiple modality significantly improves the predicting performance of single modality. The results suggest that integrating subjects with missing either CDEs/biomarker or CT imaging features can contribute to the discriminatory power of our model (p < 0.05, bootstrap two-tailed test). In summary, the proposed M3Net framework provides an effective way to integrate image and non-image data in the context of missing information.
临床数据元素(CDEs)(如年龄、吸烟史)、血液标志物和胸部计算机断层扫描(CT)结构特征已被视为评估肺癌风险的有效手段。这些自变量可以提供互补信息,我们假设将它们结合起来会提高预测准确性。在实际中,并非所有患者都具备所有这些变量。在本文中,我们提出了一种新的网络设计,称为多路径多模态缺失网络(M3Net),以使用多路径神经网络整合考虑缺失模态的多模态数据(即CDEs、生物标志物和CT图像)。每条路径学习一种模态的判别特征,不同模态在第二阶段融合以进行综合预测。该网络可以使用医学图像特征和CDEs/生物标志物进行端到端训练,也可以使用单模态进行预测。我们使用来自筛查发现病变分子和细胞特征联盟(MCL)项目三个站点的数据集对M3Net进行评估。我们的方法在1291名受试者(383名具有完整CDEs/生物标志物和CT图像)的队列中进行交叉验证,并在99名受试者(99名具有完整CDEs/生物标志物和CT图像)的队列中进行外部验证。交叉验证和外部验证结果均表明,结合多种模态可显著提高单模态的预测性能。结果表明,整合缺失CDEs/生物标志物或CT成像特征的受试者有助于提高我们模型的判别能力(p<0.05,自展双尾检验)。总之,所提出的M3Net框架提供了一种在缺失信息情况下整合图像和非图像数据的有效方法。