Erol Tugberk, Sarikaya Duygu
Computer Engineering Graduate School of Natural and Applied Sciences Gazi University Ankara Türkiye.
School of Computer Science University of Leeds Leeds United Kingdom.
Healthc Technol Lett. 2024 Dec 13;11(6):365-373. doi: 10.1049/htl2.12105. eCollection 2024 Dec.
Deep learning models are used to minimize the number of polyps that goes unnoticed by the experts and to accurately segment the detected polyps during interventions. Although state-of-the-art models are proposed, it remains a challenge to define representations that are able to generalize well and that mediate between capturing low-level features and higher-level semantic details without being redundant. Another challenge with these models is that they are computation and memory intensive, which can pose a problem with real-time applications. To address these problems, PlutoNet is proposed for polyp segmentation which requires only 9 FLOPs and 2,626,537 parameters, less than 10% of the parameters required by its counterparts. With PlutoNet, a novel approach is proposed that consists of a shared encoder, the , which is a combination of the partial decoder and full-scale connections that capture salient features at different scales without redundancy, and the auxiliary decoder which focuses on higher-level semantic features. The and the auxiliary decoder are trained with a combined loss to enforce consistency, which helps strengthen learned representations. Ablation studies and experiments are performed which show that PlutoNet performs significantly better than the state-of-the-art models, particularly on unseen datasets.
深度学习模型用于尽量减少被专家忽略的息肉数量,并在干预过程中准确分割检测到的息肉。尽管已经提出了最先进的模型,但定义能够很好地泛化、在捕获低级特征和高级语义细节之间进行调解且不会冗余的表示仍然是一个挑战。这些模型的另一个挑战是它们计算量和内存需求大,这可能给实时应用带来问题。为了解决这些问题,提出了用于息肉分割的PlutoNet,它仅需要9次浮点运算和2626537个参数,不到同类模型所需参数的10%。借助PlutoNet,提出了一种新颖的方法,该方法由一个共享编码器、即由部分解码器和全尺度连接组合而成的结构组成,该结构可在不同尺度捕获显著特征而无冗余,以及专注于高级语义特征的辅助解码器。共享编码器和辅助解码器通过组合损失进行训练以增强一致性,这有助于强化学习到的表示。进行了消融研究和实验,结果表明PlutoNet的性能明显优于最先进的模型,尤其是在未见数据集上。