• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于自动驾驶的多任务环境感知方法

Multi-Task Environmental Perception Methods for Autonomous Driving.

作者信息

Liu Ri, Yang Shubin, Tang Wansha, Yuan Jie, Chan Qiqing, Yang Yunchuan

机构信息

School of Electrical and Information Engineering, Wuhan Institute of Technology, Wuhan 430205, China.

出版信息

Sensors (Basel). 2024 Aug 28;24(17):5552. doi: 10.3390/s24175552.

DOI:10.3390/s24175552
PMID:39275463
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11398217/
Abstract

In autonomous driving, environmental perception technology often encounters challenges such as false positives, missed detections, and low accuracy, particularly in detecting small objects and complex scenarios. Existing algorithms frequently suffer from issues like feature redundancy, insufficient contextual interaction, and inadequate information fusion, making it difficult to perform multi-task detection and segmentation efficiently. To address these challenges, this paper proposes an end-to-end multi-task environmental perception model named YOLO-Mg, designed to simultaneously perform traffic object detection, lane line detection, and drivable area segmentation. First, a multi-stage gated aggregation network (MogaNet) is employed during the feature extraction process to enhance contextual interaction by improving diversity in the channel dimension, thereby compensating for the limitations of feed-forward neural networks in contextual understanding. Second, to further improve the model's accuracy in detecting objects of various scales, a restructured weighted bidirectional feature pyramid network (BiFPN) is introduced, optimizing cross-level information fusion and enabling the model to handle object detection at different scales more accurately. Finally, the model is equipped with one detection head and two segmentation heads to achieve efficient multi-task environmental perception, ensuring the simultaneous execution of multiple tasks. The experimental results on the BDD100K dataset demonstrate that the model achieves a mean average precision (mAP50) of 81.4% in object detection, an Intersection over Union (IoU) of 28.9% in lane detection, and a mean Intersection over Union (mIoU) of 92.6% in drivable area segmentation. The tests conducted in real-world scenarios show that the model performs effectively, significantly enhancing environmental perception in autonomous driving and laying a solid foundation for safer and more reliable autonomous driving systems.

摘要

在自动驾驶中,环境感知技术常常面临误报、漏检和低精度等挑战,尤其是在检测小物体和复杂场景时。现有算法经常存在特征冗余、上下文交互不足和信息融合不充分等问题,难以高效地执行多任务检测和分割。为应对这些挑战,本文提出了一种名为YOLO-Mg的端到端多任务环境感知模型,旨在同时执行交通目标检测、车道线检测和可行驶区域分割。首先,在特征提取过程中采用了多级门控聚合网络(MogaNet),通过提高通道维度的多样性来增强上下文交互,从而弥补前馈神经网络在上下文理解方面的局限性。其次,为进一步提高模型在检测各种尺度物体时的准确性,引入了一种重构的加权双向特征金字塔网络(BiFPN),优化跨层信息融合,使模型能够更准确地处理不同尺度的目标检测。最后,该模型配备了一个检测头和两个分割头,以实现高效的多任务环境感知,确保多个任务同时执行。在BDD100K数据集上的实验结果表明,该模型在目标检测中的平均精度均值(mAP50)达到81.4%,在车道检测中的交并比(IoU)为28.9%,在可行驶区域分割中的平均交并比(mIoU)为92.6%。在实际场景中进行的测试表明,该模型运行有效,显著增强了自动驾驶中的环境感知能力,为更安全、更可靠的自动驾驶系统奠定了坚实基础。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/d40dab2469a4/sensors-24-05552-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/c044d4720221/sensors-24-05552-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/a58414f7d444/sensors-24-05552-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/06b5952bf521/sensors-24-05552-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/e7cfe44e2be1/sensors-24-05552-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/366115768715/sensors-24-05552-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/e5da6439fb51/sensors-24-05552-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/486d32c96312/sensors-24-05552-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/d6bb9174cbf1/sensors-24-05552-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/ed959c66b7f1/sensors-24-05552-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/2dbf750973ad/sensors-24-05552-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/f5713899f3f9/sensors-24-05552-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/58e41e495e07/sensors-24-05552-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/d40dab2469a4/sensors-24-05552-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/c044d4720221/sensors-24-05552-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/a58414f7d444/sensors-24-05552-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/06b5952bf521/sensors-24-05552-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/e7cfe44e2be1/sensors-24-05552-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/366115768715/sensors-24-05552-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/e5da6439fb51/sensors-24-05552-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/486d32c96312/sensors-24-05552-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/d6bb9174cbf1/sensors-24-05552-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/ed959c66b7f1/sensors-24-05552-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/2dbf750973ad/sensors-24-05552-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/f5713899f3f9/sensors-24-05552-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/58e41e495e07/sensors-24-05552-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ab43/11398217/d40dab2469a4/sensors-24-05552-g013.jpg

相似文献

1
Multi-Task Environmental Perception Methods for Autonomous Driving.用于自动驾驶的多任务环境感知方法
Sensors (Basel). 2024 Aug 28;24(17):5552. doi: 10.3390/s24175552.
2
A panoramic driving perception fusion algorithm based on multi-task learning.基于多任务学习的全景驾驶感知融合算法。
PLoS One. 2024 Jun 4;19(6):e0304691. doi: 10.1371/journal.pone.0304691. eCollection 2024.
3
A Multi-Task Network Based on Dual-Neck Structure for Autonomous Driving Perception.一种基于双颈部结构的多任务网络用于自动驾驶感知。
Sensors (Basel). 2024 Feb 28;24(5):1547. doi: 10.3390/s24051547.
4
Mobip: a lightweight model for driving perception using MobileNet.Mobip:一种使用MobileNet进行驾驶感知的轻量级模型。
Front Neurorobot. 2023 Dec 4;17:1291875. doi: 10.3389/fnbot.2023.1291875. eCollection 2023.
5
A Multi-Task Road Feature Extraction Network with Grouped Convolution and Attention Mechanisms.一种具有分组卷积和注意力机制的多任务道路特征提取网络。
Sensors (Basel). 2023 Sep 30;23(19):8182. doi: 10.3390/s23198182.
6
Research on Road Scene Understanding of Autonomous Vehicles Based on Multi-Task Learning.基于多任务学习的自动驾驶车辆道路场景理解研究。
Sensors (Basel). 2023 Jul 7;23(13):6238. doi: 10.3390/s23136238.
7
MineSDS: A Unified Framework for Small Object Detection and Drivable Area Segmentation for Open-Pit Mining Scenario.MineSDS:露天采矿场景中小目标检测和可行驶区域分割的统一框架。
Sensors (Basel). 2023 Jun 27;23(13):5977. doi: 10.3390/s23135977.
8
Enhanced YOLOv5: An Efficient Road Object Detection Method.增强型YOLOv5:一种高效的道路目标检测方法。
Sensors (Basel). 2023 Oct 10;23(20):8355. doi: 10.3390/s23208355.
9
MRD-YOLO: A Multispectral Object Detection Algorithm for Complex Road Scenes.MRD-YOLO:一种用于复杂道路场景的多光谱目标检测算法。
Sensors (Basel). 2024 May 18;24(10):3222. doi: 10.3390/s24103222.
10
CenterPNets: A Multi-Task Shared Network for Traffic Perception.CenterPNets:一种用于交通感知的多任务共享网络。
Sensors (Basel). 2023 Feb 23;23(5):2467. doi: 10.3390/s23052467.

引用本文的文献

1
Leaf area estimation in small-seeded broccoli using a lightweight instance segmentation framework based on improved YOLOv11-AreaNet.基于改进的YOLOv11-AreaNet轻量级实例分割框架的小籽西兰花叶片面积估计
Front Plant Sci. 2025 Jul 4;16:1622713. doi: 10.3389/fpls.2025.1622713. eCollection 2025.

本文引用的文献

1
Heterogeneous Network Representation Learning: A Unified Framework with Survey and Benchmark.异构网络表示学习:一个包含综述与基准测试的统一框架
IEEE Trans Knowl Data Eng. 2022 Oct;34(10):4854-4873. doi: 10.1109/tkde.2020.3045924. Epub 2020 Dec 21.
2
Research on Road Scene Understanding of Autonomous Vehicles Based on Multi-Task Learning.基于多任务学习的自动驾驶车辆道路场景理解研究。
Sensors (Basel). 2023 Jul 7;23(13):6238. doi: 10.3390/s23136238.
3
Ultra Fast Deep Lane Detection With Hybrid Anchor Driven Ordinal Classification.
基于混合锚点驱动有序分类的超快速深层车道检测
IEEE Trans Pattern Anal Mach Intell. 2024 May;46(5):2555-2568. doi: 10.1109/TPAMI.2022.3182097. Epub 2024 Apr 3.
4
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.更快的 R-CNN:基于区域建议网络的实时目标检测。
IEEE Trans Pattern Anal Mach Intell. 2017 Jun;39(6):1137-1149. doi: 10.1109/TPAMI.2016.2577031. Epub 2016 Jun 6.