• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

ACU-TransNet:用于息肉分割的注意力与卷积增强型UNet-Transformer网络

ACU-TransNet: Attention and convolution-augmented UNet-transformer network for polyp segmentation.

作者信息

Huang Lei, Wu Yun

机构信息

State Key Laboratory of Public Big Data, Guizhou University, Guiyang, China.

College of Computer Science and Technology, Guizhou University, Guiyang, China.

出版信息

J Xray Sci Technol. 2024;32(6):1449-1464. doi: 10.3233/XST-240076.

DOI:10.3233/XST-240076
PMID:39422983
Abstract

BACKGROUND

UNet has achieved great success in medical image segmentation. However, due to the inherent locality of convolution operations, UNet is deficient in capturing global features and long-range dependencies of polyps, resulting in less accurate polyp recognition for complex morphologies and backgrounds. Transformers, with their sequential operations, are better at perceiving global features but lack low-level details, leading to limited localization ability. If the advantages of both architectures can be effectively combined, the accuracy of polyp segmentation can be further improved.

METHODS

In this paper, we propose an attention and convolution-augmented UNet-Transformer Network (ACU-TransNet) for polyp segmentation. This network is composed of the comprehensive attention UNet and the Transformer head, sequentially connected by the bridge layer. On the one hand, the comprehensive attention UNet enhances specific feature extraction through deformable convolution and channel attention in the first layer of the encoder and achieves more accurate shape extraction through spatial attention and channel attention in the decoder. On the other hand, the Transformer head supplements fine-grained information through convolutional attention and acquires hierarchical global characteristics from the feature maps.

RESULTS

mcU-TransNet could comprehensively learn dataset features and enhance colonoscopy interpretability for polyp detection.

CONCLUSION

Experimental results on the CVC-ClinicDB and Kvasir-SEG datasets demonstrate that mcU-TransNet outperforms existing state-of-the-art methods, showcasing its robustness.

摘要

背景

U-Net在医学图像分割方面取得了巨大成功。然而,由于卷积操作固有的局部性,U-Net在捕捉息肉的全局特征和长程依赖性方面存在不足,导致对复杂形态和背景的息肉识别不够准确。Transformer通过其序列操作,更擅长感知全局特征,但缺乏低级细节,导致定位能力有限。如果能有效结合这两种架构的优势,息肉分割的准确性可以进一步提高。

方法

在本文中,我们提出了一种用于息肉分割的注意力和卷积增强U-Net-Transformer网络(ACU-TransNet)。该网络由综合注意力U-Net和Transformer头部组成,通过桥接层顺序连接。一方面,综合注意力U-Net在编码器的第一层通过可变形卷积和通道注意力增强特定特征提取,并在解码器中通过空间注意力和通道注意力实现更准确的形状提取。另一方面,Transformer头部通过卷积注意力补充细粒度信息,并从特征图中获取分层全局特征。

结果

mcU-TransNet可以全面学习数据集特征,并增强结肠镜检查对息肉检测的可解释性。

结论

在CVC-ClinicDB和Kvasir-SEG数据集上的实验结果表明,mcU-TransNet优于现有的最先进方法,展示了其鲁棒性。

相似文献

1
ACU-TransNet: Attention and convolution-augmented UNet-transformer network for polyp segmentation.ACU-TransNet:用于息肉分割的注意力与卷积增强型UNet-Transformer网络
J Xray Sci Technol. 2024;32(6):1449-1464. doi: 10.3233/XST-240076.
2
DCATNet: polyp segmentation with deformable convolution and contextual-aware attention network.DCATNet:基于可变形卷积和上下文感知注意力网络的息肉分割
BMC Med Imaging. 2025 Apr 14;25(1):120. doi: 10.1186/s12880-025-01661-w.
3
EPSegNet: Lightweight Semantic Recalibration and Assembly for Efficient Polyp Segmentation.EPSegNet:用于高效息肉分割的轻量级语义重新校准与组装
IEEE Trans Neural Netw Learn Syst. 2025 Aug;36(8):13805-13817. doi: 10.1109/TNNLS.2025.3527557.
4
VMDU-net: a dual encoder multi-scale fusion network for polyp segmentation with Vision Mamba and Cross-Shape Transformer integration.VMDU-net:一种用于息肉分割的双编码器多尺度融合网络,集成了视觉曼巴和十字形变换器
Front Artif Intell. 2025 Jun 18;8:1557508. doi: 10.3389/frai.2025.1557508. eCollection 2025.
5
Enhancing colorectal polyp segmentation with TCFMA-Net: A transformer-based cross feature and multi-attention network.使用TCFMA-Net增强结直肠息肉分割:一种基于Transformer的交叉特征和多注意力网络
Artif Intell Med. 2025 Sep;167:103167. doi: 10.1016/j.artmed.2025.103167. Epub 2025 May 22.
6
[SG-UNet: a melanoma segmentation model enhanced with global attention and self-calibrated convolution].[SG-UNet:一种通过全局注意力和自校准卷积增强的黑色素瘤分割模型]
Nan Fang Yi Ke Da Xue Xue Bao. 2025 Jun 20;45(6):1317-1326. doi: 10.12122/j.issn.1673-4254.2025.06.21.
7
CLT-MambaSeg: An integrated model of Convolution, Linear Transformer and Multiscale Mamba for medical image segmentation.CLT-MambaSeg:一种用于医学图像分割的卷积、线性变换器和多尺度曼巴的集成模型。
Comput Biol Med. 2025 Sep;196(Pt B):110736. doi: 10.1016/j.compbiomed.2025.110736. Epub 2025 Jul 26.
8
VMKLA-UNet: vision Mamba with KAN linear attention U-Net.VMKLA-UNet:带KAN线性注意力机制的视觉曼巴U-Net
Sci Rep. 2025 Apr 17;15(1):13258. doi: 10.1038/s41598-025-97397-2.
9
TLTNet: A novel transscale cascade layered transformer network for enhanced retinal blood vessel segmentation.TLTNet:一种新颖的跨尺度级联分层Transformer 网络,用于增强视网膜血管分割。
Comput Biol Med. 2024 Aug;178:108773. doi: 10.1016/j.compbiomed.2024.108773. Epub 2024 Jun 25.
10
A novel recursive transformer-based U-Net architecture for enhanced multi-scale medical image segmentation.一种基于递归变压器的新型U-Net架构,用于增强多尺度医学图像分割。
Comput Biol Med. 2025 Sep;196(Pt A):110658. doi: 10.1016/j.compbiomed.2025.110658. Epub 2025 Jul 6.