Feng Yuncong, Cong Yeming, Xing Shuaijie, Wang Hairui, Zhao Cuixing, Zhang Xiaoli, Yao Qingan
College of Computer Science and Engineering, Changchun University of Technology, Changchun 130012, China.
Artificial Intelligence Research Institute, Changchun University of Technology, Changchun 130012, China.
Entropy (Basel). 2023 Aug 5;25(8):1169. doi: 10.3390/e25081169.
The transformer-based U-Net network structure has gained popularity in the field of medical image segmentation. However, most networks overlook the impact of the distance between each patch on the encoding process. This paper proposes a novel GC-TransUnet for medical image segmentation. The key innovation is that it takes into account the relationships between patch blocks based on their distances, optimizing the encoding process in traditional transformer networks. This optimization results in improved encoding efficiency and reduced computational costs. Moreover, the proposed GC-TransUnet is combined with U-Net to accomplish the segmentation task. In the encoder part, the traditional vision transformer is replaced by the global context vision transformer (GC-VIT), eliminating the need for the CNN network while retaining skip connections for subsequent decoders. Experimental results demonstrate that the proposed algorithm achieves superior segmentation results compared to other algorithms when applied to medical images.
基于Transformer的U-Net网络结构在医学图像分割领域颇受欢迎。然而,大多数网络忽略了每个图像块之间的距离对编码过程的影响。本文提出了一种用于医学图像分割的新型GC-TransUnet。关键创新在于它基于图像块之间的距离考虑它们之间的关系,优化了传统Transformer网络中的编码过程。这种优化提高了编码效率并降低了计算成本。此外,所提出的GC-TransUnet与U-Net相结合以完成分割任务。在编码器部分,传统视觉Transformer被全局上下文视觉Transformer(GC-VIT)取代,无需CNN网络,同时为后续解码器保留跳跃连接。实验结果表明,该算法应用于医学图像时,与其他算法相比取得了更优的分割结果。