Li Ruihao, Zhou Lingxiao, Wang Yunpeng, Shan Fei, Chen Xinrong, Liu Lei
Academy for Engineering & Technology, Fudan University, Shanghai, China.
Institute of Microscale Optoelectronics, Shenzhen University, Shenzhen, China.
Quant Imaging Med Surg. 2023 Aug 1;13(8):5333-5348. doi: 10.21037/qims-23-2. Epub 2023 Jul 5.
Lung cancer is a global disease with high lethality, with early screening being considerably helpful for improving the 5-year survival rate. Multimodality features in early screening imaging are an important part of the prediction for lung adenocarcinoma, and establishing a model for adenocarcinoma diagnosis based on multimodal features is an obvious clinical need. Through our practice and investigation, we found that graph neural networks (GNNs) are excellent platforms for multimodal feature fusion, and the data can be completed using the edge-generation network. Therefore, we propose a new lung adenocarcinoma multiclassification model based on multimodal features and an edge-generation network.
According to a ratio of 80% to 20%, respectively, the dataset of 338 cases was divided into the training set and the test set through 5-fold cross-validation, and the distribution of the 2 sets was the same. First, the regions of interest (ROIs) cropped from computed tomography (CT) images were separately fed into convolutional neural networks (CNNs) and radiomics processing platforms. The results of the 2 parts were then input into a graph embedding representation network to obtain the fused feature vectors. Subsequently, a graph database based on the clinical and semantic features was established, and the data were supplemented by an edge-generation network, with the fused feature vectors being used as the input of the nodes. This enabled us to clearly understand where the information transmission of the GNN takes place and improves the interpretability of the model. Finally, the nodes were classified using GNNs.
On our dataset, the proposed method presented in this paper achieved superior results compared to traditional methods and showed some comparability with state-of-the-art methods for lung nodule classification. The results of our method are as follows: accuracy (ACC) =66.26% (±4.46%), area under the curve (AUC) =75.86% (±1.79%), F1-score =64.00% (±3.65%), and Matthews correlation coefficient (MCC) =48.40% (±5.07%). The model with the edge-generating network consistently outperformed the model without it in all aspects.
The experiments demonstrate that with appropriate data=construction methods GNNs can outperform traditional image processing methods in the field of CT-based medical image classification. Additionally, our model has higher interpretability, as it employs subjective clinical and semantic features as the data construction approach. This will help doctors better leverage human-computer interactions.
肺癌是一种具有高致死率的全球性疾病,早期筛查对提高5年生存率有很大帮助。早期筛查影像中的多模态特征是预测肺腺癌的重要组成部分,基于多模态特征建立腺癌诊断模型是明显的临床需求。通过我们的实践和调研,我们发现图神经网络(GNN)是多模态特征融合的优秀平台,并且可以使用边生成网络来完善数据。因此,我们提出了一种基于多模态特征和边生成网络的新型肺腺癌多分类模型。
按照80%与20%的比例,通过5折交叉验证将338例病例的数据集分别划分为训练集和测试集,两组的分布相同。首先,将从计算机断层扫描(CT)图像中裁剪出的感兴趣区域(ROI)分别输入卷积神经网络(CNN)和放射组学处理平台。然后将这两部分的结果输入到图嵌入表示网络中以获得融合特征向量。随后,基于临床和语义特征建立图数据库,并通过边生成网络对数据进行补充,将融合特征向量用作节点的输入。这使我们能够清楚地了解GNN的信息传递发生在哪里,并提高了模型的可解释性。最后,使用GNN对节点进行分类。
在我们的数据集上,本文提出的方法与传统方法相比取得了更好的结果,并且在肺结节分类方面与最先进的方法具有一定的可比性。我们方法的结果如下:准确率(ACC)=66.26%(±4.46%),曲线下面积(AUC)=75.86%(±1.79%),F1分数=64.00%(±3.65%),以及马修斯相关系数(MCC)=48.40%(±5.07%)。具有边生成网络的模型在各个方面始终优于没有边生成网络的模型。
实验表明,通过适当的数据构建方法,GNN在基于CT的医学图像分类领域可以优于传统图像处理方法。此外,我们的模型具有更高的可解释性,因为它采用主观临床和语义特征作为数据构建方法。这将有助于医生更好地利用人机交互。