• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于注意融合的单阶段多光谱行人检测

Attention Fusion for One-Stage Multispectral Pedestrian Detection.

机构信息

School of Artificial Intelligence, Beijing University of Posts and Telecommunications, Beijing 100876, China.

China Mobile Research Institute, Beijing 100053, China.

出版信息

Sensors (Basel). 2021 Jun 18;21(12):4184. doi: 10.3390/s21124184.

DOI:10.3390/s21124184
PMID:34207183
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8235776/
Abstract

Multispectral pedestrian detection, which consists of a color stream and thermal stream, is essential under conditions of insufficient illumination because the fusion of the two streams can provide complementary information for detecting pedestrians based on deep convolutional neural networks (CNNs). In this paper, we introduced and adapted a simple and efficient one-stage YOLOv4 to replace the current state-of-the-art two-stage fast-RCNN for multispectral pedestrian detection and to directly predict bounding boxes with confidence scores. To further improve the detection performance, we analyzed the existing multispectral fusion methods and proposed a novel multispectral channel feature fusion (MCFF) module for integrating the features from the color and thermal streams according to the illumination conditions. Moreover, several fusion architectures, such as Early Fusion, Halfway Fusion, Late Fusion, and Direct Fusion, were carefully designed based on the MCFF to transfer the feature information from the bottom to the top at different stages. Finally, the experimental results on the KAIST and Utokyo pedestrian benchmarks showed that Halfway Fusion was used to obtain the best performance of all architectures and the MCFF could adapt fused features in the two modalities. The log-average miss rate (MR) for the two modalities with reasonable settings were 4.91% and 23.14%, respectively.

摘要

多光谱行人检测由颜色流和热流组成,在光照条件不足的情况下至关重要,因为这两种流的融合可以为基于深度卷积神经网络(CNN)的行人检测提供互补信息。在本文中,我们引入并改编了一个简单而高效的单阶段 YOLOv4 来替代当前最先进的两阶段快速 R-CNN 进行多光谱行人检测,并直接用置信分数预测边界框。为了进一步提高检测性能,我们分析了现有的多光谱融合方法,并提出了一种新颖的多光谱通道特征融合(MCFF)模块,根据光照条件融合颜色流和热流的特征。此外,还根据 MCFF 精心设计了几种融合架构,如早期融合、中途融合、晚期融合和直接融合,以便在不同阶段从底部到顶部传递特征信息。最后,在 KAIST 和 Utokyo 行人基准上的实验结果表明,中途融合用于获得所有架构的最佳性能,并且 MCFF 可以适应两种模态的融合特征。两种模态的对数平均漏报率(MR)在合理设置下分别为 4.91%和 23.14%。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/3b60da3991e8/sensors-21-04184-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/70c415acdbb9/sensors-21-04184-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/380d884a30ff/sensors-21-04184-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/34d7eae7e5a6/sensors-21-04184-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/24fcd0739dd6/sensors-21-04184-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/df9c2d76153e/sensors-21-04184-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/76fa1130e146/sensors-21-04184-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/021b46ae1a74/sensors-21-04184-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/75ef64f769b8/sensors-21-04184-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/3b60da3991e8/sensors-21-04184-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/70c415acdbb9/sensors-21-04184-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/380d884a30ff/sensors-21-04184-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/34d7eae7e5a6/sensors-21-04184-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/24fcd0739dd6/sensors-21-04184-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/df9c2d76153e/sensors-21-04184-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/76fa1130e146/sensors-21-04184-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/021b46ae1a74/sensors-21-04184-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/75ef64f769b8/sensors-21-04184-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/88cb/8235776/3b60da3991e8/sensors-21-04184-g009.jpg

相似文献

1
Attention Fusion for One-Stage Multispectral Pedestrian Detection.基于注意融合的单阶段多光谱行人检测
Sensors (Basel). 2021 Jun 18;21(12):4184. doi: 10.3390/s21124184.
2
Adopting the YOLOv4 Architecture for Low-Latency Multispectral Pedestrian Detection in Autonomous Driving.采用 YOLOv4 架构实现自动驾驶中的低延迟多光谱行人检测。
Sensors (Basel). 2022 Jan 30;22(3):1082. doi: 10.3390/s22031082.
3
Exploiting fusion architectures for multispectral pedestrian detection and segmentation.利用融合架构进行多光谱行人检测与分割。
Appl Opt. 2018 Jun 20;57(18):D108-D116. doi: 10.1364/AO.57.00D108.
4
Pedestrian Detection Using Multispectral Images and a Deep Neural Network.基于多光谱图像和深度神经网络的行人检测。
Sensors (Basel). 2021 Apr 4;21(7):2536. doi: 10.3390/s21072536.
5
Multispectral image fusion based pedestrian detection using a multilayer fused deconvolutional single-shot detector.基于多光谱图像融合的行人检测,使用多层融合反卷积单发检测器。
J Opt Soc Am A Opt Image Sci Vis. 2020 May 1;37(5):768-779. doi: 10.1364/JOSAA.386410.
6
INSANet: INtra-INter Spectral Attention Network for Effective Feature Fusion of Multispectral Pedestrian Detection.INSANet:用于多光谱行人检测有效特征融合的内部-外部光谱注意力网络
Sensors (Basel). 2024 Feb 10;24(4):1168. doi: 10.3390/s24041168.
7
An Unsupervised Transfer Learning Framework for Visible-Thermal Pedestrian Detection.基于无监督迁移学习的可见光-热行人检测框架。
Sensors (Basel). 2022 Jun 10;22(12):4416. doi: 10.3390/s22124416.
8
TFDet: Target-Aware Fusion for RGB-T Pedestrian Detection.TFDet:用于RGB-T行人检测的目标感知融合
IEEE Trans Neural Netw Learn Syst. 2024 Aug 23;PP. doi: 10.1109/TNNLS.2024.3443455.
9
Deep Visible and Thermal Image Fusion for Enhanced Pedestrian Visibility.深度可见光与热图像融合增强行人可见度。
Sensors (Basel). 2019 Aug 28;19(17):3727. doi: 10.3390/s19173727.
10
Towards High Accuracy Pedestrian Detection on Edge GPUs.面向边缘 GPU 的高精度行人检测。
Sensors (Basel). 2022 Aug 10;22(16):5980. doi: 10.3390/s22165980.

引用本文的文献

1
MRD-YOLO: A Multispectral Object Detection Algorithm for Complex Road Scenes.MRD-YOLO:一种用于复杂道路场景的多光谱目标检测算法。
Sensors (Basel). 2024 May 18;24(10):3222. doi: 10.3390/s24103222.
2
Object Detection, Recognition, and Tracking Algorithms for ADASs-A Study on Recent Trends.用于高级驾驶辅助系统的目标检测、识别和跟踪算法——近期趋势研究
Sensors (Basel). 2023 Dec 31;24(1):249. doi: 10.3390/s24010249.
3
Node-Loss Detection Methods for CZ Silicon Single Crystal Based on Multimodal Data Fusion.基于多模态数据融合的 CZ 硅单晶体节点损耗检测方法。

本文引用的文献

1
Pedestrian Detection Using Multispectral Images and a Deep Neural Network.基于多光谱图像和深度神经网络的行人检测。
Sensors (Basel). 2021 Apr 4;21(7):2536. doi: 10.3390/s21072536.
2
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.更快的 R-CNN:基于区域建议网络的实时目标检测。
IEEE Trans Pattern Anal Mach Intell. 2017 Jun;39(6):1137-1149. doi: 10.1109/TPAMI.2016.2577031. Epub 2016 Jun 6.
3
Fast Feature Pyramids for Object Detection.快速目标检测特征金字塔。
Sensors (Basel). 2023 Jun 24;23(13):5855. doi: 10.3390/s23135855.
4
Multispectral Benchmark Dataset and Baseline for Forklift Collision Avoidance.多光谱基准数据集和叉车防撞基线。
Sensors (Basel). 2022 Oct 19;22(20):7953. doi: 10.3390/s22207953.
5
Adopting the YOLOv4 Architecture for Low-Latency Multispectral Pedestrian Detection in Autonomous Driving.采用 YOLOv4 架构实现自动驾驶中的低延迟多光谱行人检测。
Sensors (Basel). 2022 Jan 30;22(3):1082. doi: 10.3390/s22031082.
IEEE Trans Pattern Anal Mach Intell. 2014 Aug;36(8):1532-45. doi: 10.1109/TPAMI.2014.2300479.
4
Pedestrian detection: an evaluation of the state of the art.行人检测:现状评估。
IEEE Trans Pattern Anal Mach Intell. 2012 Apr;34(4):743-61. doi: 10.1109/TPAMI.2011.155.
5
Survey of pedestrian detection for advanced driver assistance systems.高级驾驶员辅助系统中的行人检测综述。
IEEE Trans Pattern Anal Mach Intell. 2010 Jul;32(7):1239-58. doi: 10.1109/TPAMI.2009.122.