• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

MHS-VIT:用于交通图像检测的曼巴混合自注意力视觉Transformer

MHS-VIT: Mamba hybrid self-attention vision transformers for traffic image detection.

作者信息

Zhang Xude, Ou Weihua, Wu Xiaoping, Zhang Changzhen

机构信息

Engineering Research Center of Micro-Nano and Intelligent Manufacturing, Ministry of Education, Kaili University, Kaili, Guizhou, China.

College of Microelectronics and Artificial Intelligence, Kaili University, Kaili, Guizhou, China.

出版信息

PLoS One. 2025 Jun 30;20(6):e0325962. doi: 10.1371/journal.pone.0325962. eCollection 2025.

DOI:10.1371/journal.pone.0325962
PMID:40587465
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12208429/
Abstract

With the rapid development of intelligent transportation systems, especially in traffic image detection tasks, the introduction of the transformer architecture greatly promotes the improvement of model performance. However, traditional transformer models have high computational costs during training and deployment due to the quadratic complexity of their self-attention mechanism, which limits their application in resource-constrained environments. To overcome this limitation, this paper proposes a novel hybrid architecture, Mamba Hybrid Self-Attention Vision Transformers (MHS-VIT), which combines the advantages of Mamba state-space model (SSM) and transformer to improve the modeling efficiency and performance of visual tasks and to enhance the modeling efficiency and accuracy of the model in processing traffic images. Mamba, as a linear time complexity SSM, can effectively reduce the computational burden without sacrificing performance. The self-attention mechanism of the transformer is good at capturing long-distance spatial dependencies in images, which is crucial for understanding complex traffic scenes. Experimental results showed that MHS-VIT exhibited excellent performances in traffic image detection tasks. Whether it is vehicle detection, pedestrian detection, or traffic sign recognition tasks, this model could accurately and quickly identify target objects. Compared with backbone networks of the same scale, MHS-VIT achieved significant improvements in accuracy and model parameter quantity.

摘要

随着智能交通系统的快速发展,尤其是在交通图像检测任务中,Transformer架构的引入极大地推动了模型性能的提升。然而,传统的Transformer模型由于其自注意力机制的二次复杂性,在训练和部署过程中计算成本较高,这限制了它们在资源受限环境中的应用。为了克服这一限制,本文提出了一种新颖的混合架构,即曼巴混合自注意力视觉Transformer(MHS-VIT),它结合了曼巴状态空间模型(SSM)和Transformer的优点,以提高视觉任务的建模效率和性能,并增强模型在处理交通图像时的建模效率和准确性。曼巴作为一种具有线性时间复杂度的SSM,可以在不牺牲性能的情况下有效减轻计算负担。Transformer的自注意力机制擅长捕捉图像中的长距离空间依赖性,这对于理解复杂的交通场景至关重要。实验结果表明,MHS-VIT在交通图像检测任务中表现出色。无论是车辆检测、行人检测还是交通标志识别任务,该模型都能准确快速地识别目标物体。与相同规模的骨干网络相比,MHS-VIT在准确性和模型参数量方面都取得了显著提升。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/7226e0d629aa/pone.0325962.g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/4b66ed37f20a/pone.0325962.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/d6632c0e8aed/pone.0325962.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/8af985e7d99e/pone.0325962.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/8e754718a10b/pone.0325962.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/26d1375e4150/pone.0325962.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/97ba6aed6a1f/pone.0325962.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/c67366c80840/pone.0325962.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/c3a8838cbed1/pone.0325962.g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/7226e0d629aa/pone.0325962.g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/4b66ed37f20a/pone.0325962.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/d6632c0e8aed/pone.0325962.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/8af985e7d99e/pone.0325962.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/8e754718a10b/pone.0325962.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/26d1375e4150/pone.0325962.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/97ba6aed6a1f/pone.0325962.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/c67366c80840/pone.0325962.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/c3a8838cbed1/pone.0325962.g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c163/12208429/7226e0d629aa/pone.0325962.g009.jpg

相似文献

1
MHS-VIT: Mamba hybrid self-attention vision transformers for traffic image detection.MHS-VIT:用于交通图像检测的曼巴混合自注意力视觉Transformer
PLoS One. 2025 Jun 30;20(6):e0325962. doi: 10.1371/journal.pone.0325962. eCollection 2025.
2
Multicenter Histology Image Integration and Multiscale Deep Learning for Machine Learning-Enabled Pediatric Sarcoma Classification.用于支持机器学习的小儿肉瘤分类的多中心组织学图像整合与多尺度深度学习
medRxiv. 2025 Jun 11:2025.06.10.25328700. doi: 10.1101/2025.06.10.25328700.
3
VMDU-net: a dual encoder multi-scale fusion network for polyp segmentation with Vision Mamba and Cross-Shape Transformer integration.VMDU-net:一种用于息肉分割的双编码器多尺度融合网络,集成了视觉曼巴和十字形变换器
Front Artif Intell. 2025 Jun 18;8:1557508. doi: 10.3389/frai.2025.1557508. eCollection 2025.
4
Selective State Space Models Outperform Transformers at Predicting RNA-Seq Read Coverage.在预测RNA测序读段覆盖度方面,选择性状态空间模型优于Transformer模型。
bioRxiv. 2025 Feb 17:2025.02.13.638190. doi: 10.1101/2025.02.13.638190.
5
CDA-mamba: cross-directional attention mamba for enhanced 3D medical image segmentation.CDA-曼巴:用于增强3D医学图像分割的交叉方向注意力曼巴
Sci Rep. 2025 Jul 1;15(1):21357. doi: 10.1038/s41598-025-06462-3.
6
Management of urinary stones by experts in stone disease (ESD 2025).结石病专家对尿路结石的管理(2025年结石病专家共识)
Arch Ital Urol Androl. 2025 Jun 30;97(2):14085. doi: 10.4081/aiua.2025.14085.
7
Transformers for Neuroimage Segmentation: Scoping Review.用于神经图像分割的变压器:范围综述。
J Med Internet Res. 2025 Jan 29;27:e57723. doi: 10.2196/57723.
8
Two-stage Mamba-based diffusion model for image restoration.用于图像恢复的基于曼巴的两阶段扩散模型。
Sci Rep. 2025 Jul 1;15(1):22265. doi: 10.1038/s41598-025-07032-3.
9
A deep learning approach to direct immunofluorescence pattern recognition in autoimmune bullous diseases.深度学习方法在自身免疫性大疱性疾病中的直接免疫荧光模式识别。
Br J Dermatol. 2024 Jul 16;191(2):261-266. doi: 10.1093/bjd/ljae142.
10
Enhanced Maize Leaf Disease Detection and Classification Using an Integrated CNN-ViT Model.使用集成的卷积神经网络-视觉Transformer模型增强玉米叶部病害检测与分类
Food Sci Nutr. 2025 Jun 30;13(7):e70513. doi: 10.1002/fsn3.70513. eCollection 2025 Jul.

本文引用的文献

1
An application of YOLOv8 integrated with attention mechanisms for detection of grape leaf black rot spots.集成注意力机制的YOLOv8在葡萄叶黑腐病斑点检测中的应用
PLoS One. 2025 Apr 15;20(4):e0321788. doi: 10.1371/journal.pone.0321788. eCollection 2025.
2
TransVOD: End-to-End Video Object Detection With Spatial-Temporal Transformers.TransVOD:基于时空变换的端到端视频目标检测
IEEE Trans Pattern Anal Mach Intell. 2023 Jun;45(6):7853-7869. doi: 10.1109/TPAMI.2022.3223955. Epub 2023 May 5.
3
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.
更快的 R-CNN:基于区域建议网络的实时目标检测。
IEEE Trans Pattern Anal Mach Intell. 2017 Jun;39(6):1137-1149. doi: 10.1109/TPAMI.2016.2577031. Epub 2016 Jun 6.