Suppr超能文献

XBound-Former:面向 Transformer 的跨尺度边界建模。

XBound-Former: Toward Cross-Scale Boundary Modeling in Transformers.

出版信息

IEEE Trans Med Imaging. 2023 Jun;42(6):1735-1745. doi: 10.1109/TMI.2023.3236037. Epub 2023 Jun 1.

Abstract

Skin lesion segmentation from dermoscopy images is of great significance in the quantitative analysis of skin cancers, which is yet challenging even for dermatologists due to the inherent issues, i.e., considerable size, shape and color variation, and ambiguous boundaries. Recent vision transformers have shown promising performance in handling the variation through global context modeling. Still, they have not thoroughly solved the problem of ambiguous boundaries as they ignore the complementary usage of the boundary knowledge and global contexts. In this paper, we propose a novel cross-scale boundary-aware transformer, XBound-Former, to simultaneously address the variation and boundary problems of skin lesion segmentation. XBound-Former is a purely attention-based network and catches boundary knowledge via three specially designed learners. First, we propose an implicit boundary learner (im-Bound) to constrain the network attention on the points with noticeable boundary variation, enhancing the local context modeling while maintaining the global context. Second, we propose an explicit boundary learner (ex-Bound) to extract the boundary knowledge at multiple scales and convert it into embeddings explicitly. Third, based on the learned multi-scale boundary embeddings, we propose a cross-scale boundary learner (X-Bound) to simultaneously address the problem of ambiguous and multi-scale boundaries by using learned boundary embedding from one scale to guide the boundary-aware attention on the other scales. We evaluate the model on two skin lesion datasets and one polyp lesion dataset, where our model consistently outperforms other convolution- and transformer-based models, especially on the boundary-wise metrics. All resources could be found in https://github.com/jcwang123/xboundformer.

摘要

皮肤病变分割从皮肤镜图像在皮肤癌的定量分析中具有重要意义,即使对于皮肤科医生来说,由于固有的问题,如相当大的大小、形状和颜色变化,以及不明确的边界,这也是具有挑战性的。最近的视觉转换器在通过全局上下文建模处理变化方面表现出了很有前景的性能。尽管如此,它们并没有彻底解决边界不明确的问题,因为它们忽略了边界知识和全局上下文的互补使用。在本文中,我们提出了一种新的跨尺度边界感知转换器 XBound-Former,以同时解决皮肤病变分割的变化和边界问题。XBound-Former 是一个纯粹基于注意力的网络,通过三个专门设计的学习者捕捉边界知识。首先,我们提出了一个隐式边界学习者(im-Bound),通过在具有明显边界变化的点上约束网络注意力,增强局部上下文建模,同时保持全局上下文。其次,我们提出了一个显式边界学习者(ex-Bound),以在多个尺度上提取边界知识,并显式地将其转换为嵌入。第三,基于学习到的多尺度边界嵌入,我们提出了一个跨尺度边界学习者(X-Bound),通过使用从一个尺度学习到的边界嵌入来指导另一个尺度的边界感知注意力,同时解决模糊和多尺度边界的问题。我们在两个皮肤病变数据集和一个息肉病变数据集上评估了该模型,我们的模型在所有指标上都优于其他卷积和基于转换器的模型,尤其是在边界指标上。所有资源都可以在 https://github.com/jcwang123/xboundformer 中找到。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验