Department of Computer Science and Technology, Shantou University, Shantou, China; School of Computer Science and Technology, Heilongjiang University, Harbin, China.
School of Computer Science and Technology, Heilongjiang University, Harbin, China.
Comput Biol Med. 2024 Jul;177:108640. doi: 10.1016/j.compbiomed.2024.108640. Epub 2024 May 21.
Graph convolutional neural networks (GCN) have shown the promise in medical image segmentation due to the flexibility of representing diverse range of image regions using graph nodes and propagating knowledge via graph edges. However, existing methods did not fully exploit the various attributes of image nodes and the context relationship among their attributes. We propose a new segmentation method with multi-similarity view enhancement and node attribute context learning (MNSeg). First, multiple views were formed by measuring the similarities among the image nodes, and MNSeg has a GCN based multi-view image node attribute learning (MAL) module to integrate various node attributes learnt from multiple similarity views. Each similarity view contains the specific similarities among all the image nodes, and it was integrated with the node attributes from all the channels to form the enhanced attributes of image nodes. Second, the context relationships among the attributes of image nodes are formulated by a transformer-based context relationship encoding (CRE) strategy to propagate these relationships across all the image nodes. During the transformer-based learning, the relationships were estimated based on the self-attention on all the image nodes, and then they were encoded into the learned node features. Finally, we design an attention at attribute category level (ACA) to discriminate and fuse the learnt diverse information from MAL, CRE, and the original node attributes. ACA identifies the more informative attribute categories by adaptively learn their importance. We validate the performance of MNSeg on a public lung tumor CT dataset and an in-house non-small cell lung cancer (NSCLC) dataset collected from the hospital. The segmentation results show that MNSeg outperformed the compared segmentation methods in terms of spatial overlap and the shape similarities. The ablation studies demonstrated the effectiveness of MAL, CRE, and ACA. The generalization ability of MNSeg was proved by the consistent improved segmentation performances using different 3D segmentation backbones.
图卷积神经网络(GCN)在医学图像分割中表现出了很好的性能,因为它可以使用图节点来表示各种范围的图像区域,并通过图边缘传播知识。然而,现有的方法并没有充分利用图像节点的各种属性和它们属性之间的上下文关系。我们提出了一种新的分割方法,名为多相似视图增强和节点属性上下文学习(MNSeg)。首先,通过测量图像节点之间的相似性形成多个视图,MNSeg 有一个基于 GCN 的多视图图像节点属性学习(MAL)模块,用于整合从多个相似视图中学习到的各种节点属性。每个相似视图都包含所有图像节点之间的特定相似性,它与来自所有通道的节点属性集成,形成图像节点的增强属性。其次,基于变压器的上下文关系编码(CRE)策略来制定图像节点属性之间的上下文关系,以在所有图像节点之间传播这些关系。在基于变压器的学习中,基于所有图像节点的自注意力来估计这些关系,然后将它们编码到学习到的节点特征中。最后,我们设计了一个在属性类别水平上的注意力(ACA),以区分和融合从 MAL、CRE 和原始节点属性中学到的多样化信息。ACA 通过自适应地学习它们的重要性来识别更具信息量的属性类别。我们在一个公共的肺肿瘤 CT 数据集和一个来自医院的内部非小细胞肺癌(NSCLC)数据集上验证了 MNSeg 的性能。分割结果表明,MNSeg 在空间重叠和形状相似性方面优于比较的分割方法。消融研究证明了 MAL、CRE 和 ACA 的有效性。MNSeg 的泛化能力通过使用不同的 3D 分割骨干网络获得的一致提高的分割性能得到了证明。