• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

PolypNextLSTM:一种使用 ConvNext 和 ConvLSTM 的轻量级快速息肉视频分割网络。

PolypNextLSTM: a lightweight and fast polyp video segmentation network using ConvNext and ConvLSTM.

机构信息

Institute of Medical Technology and Intelligent Systems, Technische Universitaet Hamburg, Hamburg, Germany.

出版信息

Int J Comput Assist Radiol Surg. 2024 Oct;19(10):2111-2119. doi: 10.1007/s11548-024-03244-6. Epub 2024 Aug 8.

DOI:10.1007/s11548-024-03244-6
PMID:39115609
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11442634/
Abstract

PURPOSE

Commonly employed in polyp segmentation, single-image UNet architectures lack the temporal insight clinicians gain from video data in diagnosing polyps. To mirror clinical practices more faithfully, our proposed solution, PolypNextLSTM, leverages video-based deep learning, harnessing temporal information for superior segmentation performance with least parameter overhead, making it possibly suitable for edge devices.

METHODS

PolypNextLSTM employs a UNet-like structure with ConvNext-Tiny as its backbone, strategically omitting the last two layers to reduce parameter overhead. Our temporal fusion module, a Convolutional Long Short Term Memory (ConvLSTM), effectively exploits temporal features. Our primary novelty lies in PolypNextLSTM, which stands out as the leanest in parameters and the fastest model, surpassing the performance of five state-of-the-art image and video-based deep learning models. The evaluation of the SUN-SEG dataset spans easy-to-detect and hard-to-detect polyp scenarios, along with videos containing challenging artefacts like fast motion and occlusion.

RESULTS

Comparison against 5 image-based and 5 video-based models demonstrates PolypNextLSTM's superiority, achieving a Dice score of 0.7898 on the hard-to-detect polyp test set, surpassing image-based PraNet (0.7519) and video-based PNS+ (0.7486). Notably, our model excels in videos featuring complex artefacts such as ghosting and occlusion.

CONCLUSION

PolypNextLSTM, integrating pruned ConvNext-Tiny with ConvLSTM for temporal fusion, not only exhibits superior segmentation performance but also maintains the highest frames per speed among evaluated models. Code can be found here: https://github.com/mtec-tuhh/PolypNextLSTM .

摘要

目的

常用于息肉分割,单图像 UNet 架构缺乏临床医生从视频数据中诊断息肉获得的时间洞察力。为了更忠实地反映临床实践,我们提出的解决方案 PolypNextLSTM 利用基于视频的深度学习,利用时间信息实现卓越的分割性能,同时参数开销最小,使其可能适用于边缘设备。

方法

PolypNextLSTM 采用类似于 UNet 的结构,以 ConvNext-Tiny 作为其骨干,策略性地省略最后两层以减少参数开销。我们的时间融合模块是一个卷积长短期记忆(ConvLSTM),有效地利用了时间特征。我们的主要创新在于 PolypNextLSTM,它是参数最精简、速度最快的模型,超过了五个最先进的基于图像和视频的深度学习模型的性能。对 SUN-SEG 数据集的评估涵盖了易于检测和难以检测的息肉场景,以及包含快速运动和遮挡等挑战性伪影的视频。

结果

与 5 个基于图像和 5 个基于视频的模型进行比较,证明了 PolypNextLSTM 的优越性,在难以检测的息肉测试集上达到了 0.7898 的骰子分数,超过了基于图像的 PraNet(0.7519)和基于视频的 PNS+(0.7486)。值得注意的是,我们的模型在具有复杂伪影(如重影和遮挡)的视频中表现出色。

结论

PolypNextLSTM 将修剪后的 ConvNext-Tiny 与 ConvLSTM 集成进行时间融合,不仅表现出卓越的分割性能,而且在评估的模型中保持最高的帧率。代码可以在这里找到:https://github.com/mtec-tuhh/PolypNextLSTM。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c406/11442634/56849963c1ad/11548_2024_3244_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c406/11442634/b6c59d3bc9cd/11548_2024_3244_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c406/11442634/099adf5f95a8/11548_2024_3244_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c406/11442634/56849963c1ad/11548_2024_3244_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c406/11442634/b6c59d3bc9cd/11548_2024_3244_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c406/11442634/099adf5f95a8/11548_2024_3244_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c406/11442634/56849963c1ad/11548_2024_3244_Fig3_HTML.jpg

相似文献

1
PolypNextLSTM: a lightweight and fast polyp video segmentation network using ConvNext and ConvLSTM.PolypNextLSTM:一种使用 ConvNext 和 ConvLSTM 的轻量级快速息肉视频分割网络。
Int J Comput Assist Radiol Surg. 2024 Oct;19(10):2111-2119. doi: 10.1007/s11548-024-03244-6. Epub 2024 Aug 8.
2
Iterative feedback-based models for image and video polyp segmentation.基于迭代反馈的图像和视频息肉分割模型。
Comput Biol Med. 2024 Jul;177:108569. doi: 10.1016/j.compbiomed.2024.108569. Epub 2024 May 11.
3
Multi-scale nested UNet with transformer for colorectal polyp segmentation.多尺度嵌套 UNet 与 Transformer 相结合的结直肠息肉分割方法。
J Appl Clin Med Phys. 2024 Jun;25(6):e14351. doi: 10.1002/acm2.14351. Epub 2024 Mar 29.
4
NA-segformer: A multi-level transformer model based on neighborhood attention for colonoscopic polyp segmentation.NA-segformer:一种基于邻域注意力的多层次 Transformer 模型,用于结肠镜下息肉分割。
Sci Rep. 2024 Sep 28;14(1):22527. doi: 10.1038/s41598-024-74123-y.
5
Improving Automatic Polyp Detection Using CNN by Exploiting Temporal Dependency in Colonoscopy Video.利用结肠镜视频中的时间相关性来提高 CNN 自动检测息肉的性能。
IEEE J Biomed Health Inform. 2020 Jan;24(1):180-193. doi: 10.1109/JBHI.2019.2907434. Epub 2019 Apr 1.
6
Dual-branch multi-information aggregation network with transformer and convolution for polyp segmentation.基于 Transformer 和卷积的双通道多信息聚合网络用于息肉分割。
Comput Biol Med. 2024 Jan;168:107760. doi: 10.1016/j.compbiomed.2023.107760. Epub 2023 Nov 30.
7
CTNet: Contrastive Transformer Network for Polyp Segmentation.CTNet:用于息肉分割的对比 Transformer 网络。
IEEE Trans Cybern. 2024 Sep;54(9):5040-5053. doi: 10.1109/TCYB.2024.3368154. Epub 2024 Aug 26.
8
Automated polyp segmentation for colonoscopy images: A method based on convolutional neural networks and ensemble learning.结肠镜图像的自动息肉分割:一种基于卷积神经网络和集成学习的方法。
Med Phys. 2019 Dec;46(12):5666-5676. doi: 10.1002/mp.13865. Epub 2019 Oct 31.
9
Integrating Online and Offline Three-Dimensional Deep Learning for Automated Polyp Detection in Colonoscopy Videos.整合在线和离线三维深度学习用于结肠镜检查视频中的息肉自动检测
IEEE J Biomed Health Inform. 2017 Jan;21(1):65-75. doi: 10.1109/JBHI.2016.2637004. Epub 2016 Dec 7.
10
WDFF-Net: Weighted Dual-Branch Feature Fusion Network for Polyp Segmentation With Object-Aware Attention Mechanism.WDFF-Net:基于目标感知注意力机制的带权双分支特征融合网络的息肉分割。
IEEE J Biomed Health Inform. 2024 Jul;28(7):4118-4131. doi: 10.1109/JBHI.2024.3381891. Epub 2024 Jul 2.

本文引用的文献

1
A multi-centre polyp detection and segmentation dataset for generalisability assessment.用于泛化能力评估的多中心息肉检测和分割数据集。
Sci Data. 2023 Feb 6;10(1):75. doi: 10.1038/s41597-023-01981-y.
2
Focus U-Net: A novel dual attention-gated CNN for polyp segmentation during colonoscopy.聚焦 U-Net:一种新颖的双注意力门控 CNN,用于结肠镜检查中的息肉分割。
Comput Biol Med. 2021 Oct;137:104815. doi: 10.1016/j.compbiomed.2021.104815. Epub 2021 Sep 2.
3
Enhanced U-Net: A Feature Enhancement Network for Polyp Segmentation.增强型U-Net:一种用于息肉分割的特征增强网络。
Proc Int Robot Vis Conf. 2021 May;2021:181-188. doi: 10.1109/crv52889.2021.00032. Epub 2021 Jul 5.
4
Incidence and Mortality of Proximal and Distal Colorectal Cancer in Germany—Trends in the Era of Screening Colonoscopy.德国近端和远端结直肠癌的发病率和死亡率——筛查结肠镜时代的趋势。
Dtsch Arztebl Int. 2021 Apr 23;118(16):281-287. doi: 10.3238/arztebl.m2021.0111.
5
Establishing key research questions for the implementation of artificial intelligence in colonoscopy: a modified Delphi method.确立人工智能在结肠镜检查中应用的关键研究问题:一项改良 Delphi 法研究。
Endoscopy. 2021 Sep;53(9):893-901. doi: 10.1055/a-1306-7590. Epub 2021 Jan 13.
6
Development of a computer-aided detection system for colonoscopy and a publicly accessible large colonoscopy video database (with video).开发一种用于结肠镜检查的计算机辅助检测系统和一个公开可用的大型结肠镜检查视频数据库(带视频)。
Gastrointest Endosc. 2021 Apr;93(4):960-967.e3. doi: 10.1016/j.gie.2020.07.060. Epub 2020 Jul 31.
7
UNet++: A Nested U-Net Architecture for Medical Image Segmentation.U-Net++:一种用于医学图像分割的嵌套U-Net架构。
Deep Learn Med Image Anal Multimodal Learn Clin Decis Support (2018). 2018 Sep;11045:3-11. doi: 10.1007/978-3-030-00889-5_1. Epub 2018 Sep 20.
8
The Natural History of Colorectal Polyps: Overview of Predictive Static and Dynamic Features.结直肠息肉的自然史:预测静态和动态特征概述。
Gastroenterol Clin North Am. 2018 Sep;47(3):515-536. doi: 10.1016/j.gtc.2018.04.004. Epub 2018 Jun 29.
9
A Benchmark for Endoluminal Scene Segmentation of Colonoscopy Images.用于结肠镜图像内腔场景分割的基准
J Healthc Eng. 2017;2017:4037190. doi: 10.1155/2017/4037190. Epub 2017 Jul 26.
10
DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs.DeepLab:基于深度卷积网络、空洞卷积和全连接条件随机场的语义图像分割。
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):834-848. doi: 10.1109/TPAMI.2017.2699184. Epub 2017 Apr 27.