• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过多尺度金字塔特征提取增强的VM-Unet用于从膝关节MRI中分割胫股关节组织。

VM-Unet enhanced with multi-scale pyramid feature extraction for segmentation of tibiofemoral joint tissues from knee MRI.

作者信息

Wang Xin, Fu Yupeng, Lu Huimin, Xia Yuchen, Cai Xiaodong

机构信息

College of Computer Science and Engineering, Changchun University of Technology, Changchun, China.

Information Department, Jilin Qianwei Hospital, Changchun, China.

出版信息

PLoS One. 2025 Aug 28;20(8):e0330740. doi: 10.1371/journal.pone.0330740. eCollection 2025.

DOI:10.1371/journal.pone.0330740
PMID:40875643
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12393787/
Abstract

In medical imaging diagnosis, accurate segmentation of the knee joint can help doctors better observe and diagnose lesions, thereby improving diagnostic accuracy and treatment effectiveness. Vision Mamba mainly relies on the State Space Model (SSM) for feature modeling, which excels at capturing global contextual information but cannot capture local texture features. Moreover, features of different scales are not effectively integrated, resulting in the model's weak segmentation ability on small-scale tissues (such as cartilage areas). To this end, this study proposed a novel multi-scale Vision Mamba Unet (VM-Unet) framework named MSPF-VM-Unet to perform the segmentation on the femur, tibia, femoral cartilage, and tibial cartilage in knee MRI images. The proposed MSPF-VM-Unet extends VM-Unet by introducing a designed multi-scale pyramid feature extraction network named MPSK, which synergizes multi-resolution feature extraction with channel-space attention. MPSK network enhances multi-scale local feature extraction through Selective Kernel (SK) convolution and pyramid pooling. The network merges the overall context information extracted by the Vision Mamba encoder to achieve the coordinated optimization of a multi-scale hierarchical feature fusion mechanism and global long-range dependency modeling. The results of the comparative experiments on the OAI-ZIB dataset indicate that MSPF-VM-Unet significantly improves the boundary accuracy and regional consistency of the MRI tibiofemoral joint tissue structure.

摘要

在医学影像诊断中,膝关节的精确分割有助于医生更好地观察和诊断病变,从而提高诊断准确性和治疗效果。视觉曼巴主要依靠状态空间模型(SSM)进行特征建模,该模型擅长捕捉全局上下文信息,但无法捕捉局部纹理特征。此外,不同尺度的特征没有得到有效整合,导致模型对小尺度组织(如软骨区域)的分割能力较弱。为此,本研究提出了一种名为MSPF-VM-Unet的新型多尺度视觉曼巴Unet(VM-Unet)框架,用于对膝关节MRI图像中的股骨、胫骨、股骨软骨和胫骨软骨进行分割。所提出的MSPF-VM-Unet通过引入一个名为MPSK的设计多尺度金字塔特征提取网络来扩展VM-Unet,该网络将多分辨率特征提取与通道空间注意力相结合。MPSK网络通过选择性内核(SK)卷积和金字塔池化增强多尺度局部特征提取。该网络融合了视觉曼巴编码器提取的整体上下文信息,以实现多尺度分层特征融合机制和全局长程依赖建模的协同优化。在OAI-ZIB数据集上的对比实验结果表明,MSPF-VM-Unet显著提高了MRI胫股关节组织结构的边界准确性和区域一致性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a155/12393787/83a0aa1415ca/pone.0330740.g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a155/12393787/327a531a7af2/pone.0330740.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a155/12393787/d0093dd7b974/pone.0330740.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a155/12393787/130b68558de6/pone.0330740.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a155/12393787/f064759e646d/pone.0330740.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a155/12393787/d460cd881823/pone.0330740.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a155/12393787/536a86d56c37/pone.0330740.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a155/12393787/42e8a20487f0/pone.0330740.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a155/12393787/83a0aa1415ca/pone.0330740.g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a155/12393787/327a531a7af2/pone.0330740.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a155/12393787/d0093dd7b974/pone.0330740.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a155/12393787/130b68558de6/pone.0330740.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a155/12393787/f064759e646d/pone.0330740.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a155/12393787/d460cd881823/pone.0330740.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a155/12393787/536a86d56c37/pone.0330740.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a155/12393787/42e8a20487f0/pone.0330740.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a155/12393787/83a0aa1415ca/pone.0330740.g008.jpg

相似文献

1
VM-Unet enhanced with multi-scale pyramid feature extraction for segmentation of tibiofemoral joint tissues from knee MRI.通过多尺度金字塔特征提取增强的VM-Unet用于从膝关节MRI中分割胫股关节组织。
PLoS One. 2025 Aug 28;20(8):e0330740. doi: 10.1371/journal.pone.0330740. eCollection 2025.
2
VMKLA-UNet: vision Mamba with KAN linear attention U-Net.VMKLA-UNet:带KAN线性注意力机制的视觉曼巴U-Net
Sci Rep. 2025 Apr 17;15(1):13258. doi: 10.1038/s41598-025-97397-2.
3
[SG-UNet: a melanoma segmentation model enhanced with global attention and self-calibrated convolution].[SG-UNet:一种通过全局注意力和自校准卷积增强的黑色素瘤分割模型]
Nan Fang Yi Ke Da Xue Xue Bao. 2025 Jun 20;45(6):1317-1326. doi: 10.12122/j.issn.1673-4254.2025.06.21.
4
A novel image segmentation network with multi-scale and flow-guided attention for early screening of vaginal intraepithelial neoplasia (VAIN).一种用于阴道上皮内瘤变(VAIN)早期筛查的具有多尺度和流引导注意力的新型图像分割网络。
Med Phys. 2025 Aug;52(8):e18041. doi: 10.1002/mp.18041.
5
MLAgg-UNet: Advancing Medical Image Segmentation with Efficient Transformer and Mamba-Inspired Multi-Scale Sequence.MLAgg-UNet:借助高效Transformer和受曼巴启发的多尺度序列推进医学图像分割
IEEE J Biomed Health Inform. 2025 Aug 7;PP. doi: 10.1109/JBHI.2025.3596648.
6
Multi-level channel-spatial attention and light-weight scale-fusion network (MCSLF-Net): multi-level channel-spatial attention and light-weight scale-fusion transformer for 3D brain tumor segmentation.多级通道空间注意力与轻量级尺度融合网络(MCSLF-Net):用于3D脑肿瘤分割的多级通道空间注意力与轻量级尺度融合变换器
Quant Imaging Med Surg. 2025 Jul 1;15(7):6301-6325. doi: 10.21037/qims-2025-354. Epub 2025 Jun 30.
7
A novel recursive transformer-based U-Net architecture for enhanced multi-scale medical image segmentation.一种基于递归变压器的新型U-Net架构,用于增强多尺度医学图像分割。
Comput Biol Med. 2025 Sep;196(Pt A):110658. doi: 10.1016/j.compbiomed.2025.110658. Epub 2025 Jul 6.
8
Multi-module UNet++ for colon cancer histopathological image segmentation.用于结肠癌组织病理学图像分割的多模块UNet++
Sci Rep. 2025 Aug 7;15(1):28895. doi: 10.1038/s41598-025-13636-6.
9
GMAlignNet: multi-scale lightweight brain tumor image segmentation with enhanced semantic information consistency.GMAlignNet:具有增强语义信息一致性的多尺度轻量级脑肿瘤图像分割。
Phys Med Biol. 2024 May 23;69(11). doi: 10.1088/1361-6560/ad4301.
10
FED-UNet++: An Improved Nested UNet for Hippocampus Segmentation in Alzheimer's Disease Diagnosis.FED-UNet++:一种用于阿尔茨海默病诊断中海马体分割的改进型嵌套UNet
Sensors (Basel). 2025 Aug 19;25(16):5155. doi: 10.3390/s25165155.

本文引用的文献

1
nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation.nnU-Net:一种基于深度学习的生物医学图像分割的自配置方法。
Nat Methods. 2021 Feb;18(2):203-211. doi: 10.1038/s41592-020-01008-z. Epub 2020 Dec 7.
2
UNet++: A Nested U-Net Architecture for Medical Image Segmentation.U-Net++:一种用于医学图像分割的嵌套U-Net架构。
Deep Learn Med Image Anal Multimodal Learn Clin Decis Support (2018). 2018 Sep;11045:3-11. doi: 10.1007/978-3-030-00889-5_1. Epub 2018 Sep 20.
3
Automated segmentation of knee bone and cartilage combining statistical shape knowledge and convolutional neural networks: Data from the Osteoarthritis Initiative.
基于统计形状知识和卷积神经网络的膝关节骨和软骨自动分割:来自 Osteoarthritis Initiative 的数据。
Med Image Anal. 2019 Feb;52:109-118. doi: 10.1016/j.media.2018.11.009. Epub 2018 Nov 17.
4
Deep feature learning for knee cartilage segmentation using a triplanar convolutional neural network.使用三平面卷积神经网络进行膝关节软骨分割的深度特征学习
Med Image Comput Comput Assist Interv. 2013;16(Pt 2):246-53. doi: 10.1007/978-3-642-40763-5_31.
5
In vivo T(1rho) and T(2) mapping of articular cartilage in osteoarthritis of the knee using 3 T MRI.使用3T磁共振成像对膝关节骨关节炎患者的关节软骨进行体内T(1rho)和T(2)成像
Osteoarthritis Cartilage. 2007 Jul;15(7):789-97. doi: 10.1016/j.joca.2007.01.011. Epub 2007 Feb 16.