Wu Huisi, Zhao Zebin, Zhong Jiafu, Wang Wei, Wen Zhenkun, Qin Jing
IEEE Trans Cybern. 2023 Apr;53(4):2610-2621. doi: 10.1109/TCYB.2022.3162873. Epub 2023 Mar 16.
Automatic polyp segmentation from colonoscopy videos is a prerequisite for the development of a computer-assisted colon cancer examination and diagnosis system. However, it remains a very challenging task owing to the large variation of polyps, the low contrast between polyps and background, and the blurring boundaries of polyps. More importantly, real-time performance is a necessity of this task, as it is anticipated that the segmented results can be immediately presented to the doctor during the colonoscopy intervention for his/her prompt decision and action. It is difficult to develop a model with powerful representation capability, yielding satisfactory segmentation results and, simultaneously, maintaining real-time performance. In this article, we present a novel lightweight context-aware network, namely, PolypSeg+, attempting to capture distinguishable features of polyps without increasing network complexity and sacrificing time performance. To achieve this, a set of novel lightweight techniques is developed and integrated into the proposed PolypSeg+, including an adaptive scale context (ASC) module equipped with a lightweight attention mechanism to tackle the large-scale variation of polyps, an efficient global context (EGC) module to promote the fusion of low-level and high-level features by excluding background noise and preserving boundary details, and a lightweight feature pyramid fusion (FPF) module to further refine the features extracted from the ASC and EGC. We extensively evaluate the proposed PolypSeg+ on two famous public available datasets for the polyp segmentation task: 1) Kvasir-SEG and 2) CVC-Endoscenestill. The experimental results demonstrate that our PolypSeg+ consistently outperforms other state-of-the-art networks by achieving better segmentation accuracy in much less running time. The code is available at https://github.com/szu-zzb/polypsegplus.
从结肠镜检查视频中自动进行息肉分割是计算机辅助结肠癌检查与诊断系统发展的前提条件。然而,由于息肉的巨大差异、息肉与背景之间的低对比度以及息肉边界的模糊性,这仍然是一项极具挑战性的任务。更重要的是,实时性能是这项任务的必要条件,因为预计在结肠镜检查干预过程中,分割结果能够立即呈现给医生,以便其迅速做出决策并采取行动。开发一个具有强大表示能力、能产生令人满意的分割结果并同时保持实时性能的模型是很困难的。在本文中,我们提出了一种新颖的轻量级上下文感知网络,即PolypSeg+,试图在不增加网络复杂性和不牺牲时间性能的情况下捕捉息肉的可区分特征。为实现这一目标,我们开发了一套新颖的轻量级技术并将其集成到所提出的PolypSeg+中,包括一个配备轻量级注意力机制以应对息肉大规模变化的自适应尺度上下文(ASC)模块、一个通过排除背景噪声和保留边界细节来促进低级和高级特征融合的高效全局上下文(EGC)模块,以及一个用于进一步细化从ASC和EGC中提取的特征的轻量级特征金字塔融合(FPF)模块。我们在两个著名的公开可用的息肉分割任务数据集上对所提出的PolypSeg+进行了广泛评估:1)Kvasir-SEG和2)CVC-Endoscenestill。实验结果表明,我们的PolypSeg+在运行时间少得多的情况下,通过实现更好的分割精度,始终优于其他现有最先进的网络。代码可在https://github.com/szu-zzb/polypsegplus获取。