• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于复杂环境中稳健车辆跟踪的注意力增强型StrongSORT算法

Attention-enhanced StrongSORT for robust vehicle tracking in complex environments.

作者信息

Xu Wei, Du Xiaodong, Li Ruochen, Li Bingjie, Jiao Yuhu, Xing Lei

机构信息

Shandong University of Science and Technology, College of Transportation, Qingdao, 266590, China.

出版信息

Sci Rep. 2025 May 20;15(1):17472. doi: 10.1038/s41598-025-99524-5.

DOI:10.1038/s41598-025-99524-5
PMID:40394205
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12092783/
Abstract

While multi-object tracking is critical for autonomous driving systems, traditional algorithms exhibit three fundamental limitations in complex scenarios: (1) blurred feature representation under occlusion and re-identification scenarios causing identity switches, (2) insufficient sensitivity to scale-variant targets due to fixed geometric constraints in conventional IoU-based loss functions, and (3) gradient degradation in deep convolutional layers hindering discriminative feature learning. To address these challenges, we propose AE-StrongSORT (Attention-Enhanced StrongSORT), an attention-enhanced tracking framework featuring three systematic innovations: first, the GAM-YOLO (global attention mechanism-YOLO)hybrid architecture integrates multi-scale feature fusion with a global attention mechanism (GC2f structure). This design enhances cross-dimensional feature interaction through localized channel-spatial attention gates, significantly improving occlusion-resistant feature representation (IDF1 ↑ 9.99%, IDsw ↓ 9.85%). Second, the F-EIoU loss function introduces dynamic size-dependent penalty terms and difficulty-adaptive weighting factors, effectively balancing learning priorities between small targets and normal instances. Third, the optimized CBH-Conv module employs Hardswish activation and depthwise separable convolution to mitigate gradient vanishing while maintaining real-time efficiency (achieving a 17% MOTA improvement at 213 FPS).Evaluated on the MOT-16 dataset, AE-StrongSORT demonstrates substantial improvements over the baseline StrongSORT, with 17%, 2.78%, and 9.99% gains in MOTA, HOTA, and IDF1 metrics respectively, alongside significant reductions in false/missed detections. These advances establish a novel technical pathway for robust vehicle tracking in real-world traffic scenarios characterized by coexisting challenges of scale variation, motion blur, and dense occlusion.

摘要

虽然多目标跟踪对于自动驾驶系统至关重要,但传统算法在复杂场景中存在三个基本局限性:(1)在遮挡和重新识别场景下特征表示模糊,导致身份切换;(2)由于基于传统交并比(IoU)的损失函数中固定的几何约束,对尺度变化目标的敏感性不足;(3)深度卷积层中的梯度退化阻碍了判别性特征学习。为应对这些挑战,我们提出了AE-StrongSORT(注意力增强的StrongSORT),这是一个具有三项系统性创新的注意力增强跟踪框架:首先,GAM-YOLO(全局注意力机制-YOLO)混合架构将多尺度特征融合与全局注意力机制(GC2f结构)相结合。这种设计通过局部通道-空间注意力门增强了跨维度特征交互,显著提高了抗遮挡特征表示(IDF1提升9.99%,IDsw下降9.85%)。其次,F-EIoU损失函数引入了动态的尺寸相关惩罚项和难度自适应加权因子,有效平衡了小目标和正常实例之间的学习优先级。第三,优化后的CBH-Conv模块采用Hardswish激活函数和深度可分离卷积来减轻梯度消失,同时保持实时效率(在213帧每秒时MOTA提高17%)。在MOT-16数据集上进行评估时,AE-StrongSORT相较于基线StrongSORT有显著改进,在MOTA、HOTA和IDF1指标上分别提高了17%、2.78%和9.99%,同时显著减少了误检/漏检。这些进展为在存在尺度变化、运动模糊和密集遮挡等共存挑战的现实交通场景中进行稳健的车辆跟踪建立了一条新的技术途径。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/5b64bb8fa588/41598_2025_99524_Fig16a_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/729c4a7233a7/41598_2025_99524_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/215a7434ccdc/41598_2025_99524_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/38ebf6178622/41598_2025_99524_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/3d7e0207d75e/41598_2025_99524_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/5686d5ccf934/41598_2025_99524_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/1b8817b51c9f/41598_2025_99524_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/e611fe051b6a/41598_2025_99524_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/f52b818d133f/41598_2025_99524_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/8aac552f71ef/41598_2025_99524_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/91b5a9a79315/41598_2025_99524_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/8612bb0a8d7b/41598_2025_99524_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/4caa58862987/41598_2025_99524_Fig12_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/55968a4b931e/41598_2025_99524_Fig13_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/5e904a7e9701/41598_2025_99524_Fig14_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/a701ae7a0f48/41598_2025_99524_Fig15_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/5b64bb8fa588/41598_2025_99524_Fig16a_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/729c4a7233a7/41598_2025_99524_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/215a7434ccdc/41598_2025_99524_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/38ebf6178622/41598_2025_99524_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/3d7e0207d75e/41598_2025_99524_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/5686d5ccf934/41598_2025_99524_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/1b8817b51c9f/41598_2025_99524_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/e611fe051b6a/41598_2025_99524_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/f52b818d133f/41598_2025_99524_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/8aac552f71ef/41598_2025_99524_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/91b5a9a79315/41598_2025_99524_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/8612bb0a8d7b/41598_2025_99524_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/4caa58862987/41598_2025_99524_Fig12_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/55968a4b931e/41598_2025_99524_Fig13_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/5e904a7e9701/41598_2025_99524_Fig14_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/a701ae7a0f48/41598_2025_99524_Fig15_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14b2/12092783/5b64bb8fa588/41598_2025_99524_Fig16a_HTML.jpg

相似文献

1
Attention-enhanced StrongSORT for robust vehicle tracking in complex environments.用于复杂环境中稳健车辆跟踪的注意力增强型StrongSORT算法
Sci Rep. 2025 May 20;15(1):17472. doi: 10.1038/s41598-025-99524-5.
2
A Long-Term Video Tracking Method for Group-Housed Pigs.一种用于群养猪的长期视频跟踪方法。
Animals (Basel). 2024 May 19;14(10):1505. doi: 10.3390/ani14101505.
3
Helmet-Wearing Tracking Detection Based on StrongSORT.基于 StrongSORT 的戴盔检测。
Sensors (Basel). 2023 Feb 3;23(3):1682. doi: 10.3390/s23031682.
4
Improved UAV-to-Ground Multi-Target Tracking Algorithm Based on StrongSORT.基于StrongSORT的改进型无人机对地多目标跟踪算法
Sensors (Basel). 2023 Nov 17;23(22):9239. doi: 10.3390/s23229239.
5
Research on Cattle Behavior Recognition and Multi-Object Tracking Algorithm Based on YOLO-BoT.基于YOLO-BoT的牛行为识别与多目标跟踪算法研究
Animals (Basel). 2024 Oct 17;14(20):2993. doi: 10.3390/ani14202993.
6
YOLO-MARS: An Enhanced YOLOv8n for Small Object Detection in UAV Aerial Imagery.YOLO-MARS:一种用于无人机航空图像中小目标检测的增强型YOLOv8n。
Sensors (Basel). 2025 Apr 17;25(8):2534. doi: 10.3390/s25082534.
7
STMMOT: Advancing multi-object tracking through spatiotemporal memory networks and multi-scale attention pyramids.STMMOT:通过时空记忆网络和多尺度注意力金字塔推进多目标跟踪
Neural Netw. 2023 Nov;168:363-379. doi: 10.1016/j.neunet.2023.09.047. Epub 2023 Sep 29.
8
An efficient and lightweight detection method for stranded elastic needle defects in complex industrial environments using VEE-YOLO.一种使用VEE - YOLO在复杂工业环境中检测 stranded elastic needle缺陷的高效轻量级检测方法。 (注:原文中“stranded elastic needle”不太明确具体准确含义,可能存在表述不完整或有误的情况)
Sci Rep. 2025 Jan 22;15(1):2879. doi: 10.1038/s41598-025-85721-9.
9
EnhanceCenter for improving point based tracking and rich feature representation.用于改进基于点的跟踪和丰富特征表示的增强中心。
Sci Rep. 2025 Mar 3;15(1):5214. doi: 10.1038/s41598-025-88924-2.
10
RFAG-YOLO: A Receptive Field Attention-Guided YOLO Network for Small-Object Detection in UAV Images.RFAG-YOLO:一种用于无人机图像中小目标检测的感受野注意力引导YOLO网络。
Sensors (Basel). 2025 Mar 30;25(7):2193. doi: 10.3390/s25072193.

本文引用的文献

1
Green pepper fruits counting based on improved DeepSort and optimized Yolov5s.基于改进的DeepSort和优化的Yolov5s的青椒果实计数
Front Plant Sci. 2024 Jul 16;15:1417682. doi: 10.3389/fpls.2024.1417682. eCollection 2024.
2
Enhancing Geometric Factors in Model Learning and Inference for Object Detection and Instance Segmentation.增强目标检测与实例分割模型学习与推理中的几何因素
IEEE Trans Cybern. 2022 Aug;52(8):8574-8586. doi: 10.1109/TCYB.2021.3095305. Epub 2022 Jul 19.
3
HAR-Net: Joint Learning of Hybrid Attention for Single-stage Object Detection.
HAR-Net:用于单阶段目标检测的混合注意力联合学习
IEEE Trans Image Process. 2019 Dec 11. doi: 10.1109/TIP.2019.2957850.
4
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.更快的 R-CNN:基于区域建议网络的实时目标检测。
IEEE Trans Pattern Anal Mach Intell. 2017 Jun;39(6):1137-1149. doi: 10.1109/TPAMI.2016.2577031. Epub 2016 Jun 6.