School of Software, East China Jiaotong University, Nanchang, China.
School of International, East China Jiaotong University, Nanchang, China.
J Healthc Eng. 2022 Jul 11;2022:9016401. doi: 10.1155/2022/9016401. eCollection 2022.
retinal image is a crucial window for the clinical observation of cardiovascular, cerebrovascular, or other correlated diseases. Retinal vessel segmentation is of great benefit to the clinical diagnosis. Recently, the convolutional neural network (CNN) has become a dominant method in the retinal vessel segmentation field, especially the U-shaped CNN models. However, the conventional encoder in CNN is vulnerable to noisy interference, and the long-rang relationship in fundus images has not been fully utilized. In this paper, we propose a novel model called Transformer in M-Net (TiM-Net) based on M-Net, diverse attention mechanisms, and weighted side output layers to efficaciously perform retinal vessel segmentation. First, to alleviate the effects of noise, a dual-attention mechanism based on channel and spatial is designed. Then the self-attention mechanism in Transformer is introduced into skip connection to re-encode features and model the long-range relationship explicitly. Finally, a weighted SideOut layer is proposed for better utilization of the features from each side layer. Extensive experiments are conducted on three public data sets to show the effectiveness and robustness of our TiM-Net compared with the state-of-the-art baselines. Both quantitative and qualitative results prove its clinical practicality. Moreover, variants of TiM-Net also achieve competitive performance, demonstrating its scalability and generalization ability. The code of our model is available at https://github.com/ZX-ECJTU/TiM-Net.
视网膜图像是观察心血管、脑血管或其他相关疾病的重要窗口。视网膜血管分割对临床诊断有很大的帮助。最近,卷积神经网络(CNN)已成为视网膜血管分割领域的主导方法,尤其是 U 形 CNN 模型。然而,CNN 中的传统编码器容易受到噪声干扰,眼底图像中的长程关系尚未得到充分利用。在本文中,我们提出了一种基于 M-Net 的新型模型,称为 Transformer in M-Net(TiM-Net),该模型结合了多种注意力机制和加权侧输出层,可有效地进行视网膜血管分割。首先,为了减轻噪声的影响,设计了一种基于通道和空间的双注意力机制。然后,将 Transformer 中的自注意力机制引入到跳过连接中,以重新编码特征并显式建模长程关系。最后,提出了加权 SideOut 层,以更好地利用每个侧层的特征。在三个公共数据集上进行了广泛的实验,结果表明与最先进的基线相比,我们的 TiM-Net 具有有效性和鲁棒性。定量和定性结果都证明了它的临床实用性。此外,TiM-Net 的变体也取得了有竞争力的性能,证明了它的可扩展性和泛化能力。我们模型的代码可在 https://github.com/ZX-ECJTU/TiM-Net 上获取。