• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于新型轴线表示和回归模式的行人检测。

Pedestrian Detection by Novel Axis-Line Representation and Regression Pattern.

机构信息

School of Software Engineering, South China University of Technology, Guangzhou 510006, China.

出版信息

Sensors (Basel). 2021 May 11;21(10):3312. doi: 10.3390/s21103312.

DOI:10.3390/s21103312
PMID:34064653
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8150842/
Abstract

The pattern of bounding box representation and regression has long been dominant in CNN-based pedestrian detectors. Despite the method's success, it cannot accurately represent location, and introduces unnecessary background information, while pedestrian features are mainly located in axis-line areas. Other object representations, such as corner-pairs, are not easy to obtain by regression because the corners are far from the axis-line and are greatly affected by background features. In this paper, we propose a novel detection pattern, named Axis-line Representation and Regression (ALR), for pedestrian detection in road scenes. Specifically, we design a 3-d axis-line representation for pedestrians and use it as the regression target during network training. A line-box transformation method is also proposed to fit the widely used box-annotations. Meanwhile, we explore the influence of deformable convolution base-offset on detection performance and propose a base-offset initialization strategy to further promote the gain brought by ALR. Notably, the proposed ALR pattern can be introduced into both anchor-based and anchor-free frameworks. We validate the effectiveness of ALR on the Caltech-USA and CityPersons datasets. Experimental results show that our approach outperforms the baseline significantly through simple modifications and achieves competitive accuracy with other methods without bells and whistles.

摘要

基于卷积神经网络的行人检测器长期以来一直采用边界框表示和回归的模式。尽管该方法取得了成功,但它无法准确表示位置,并且引入了不必要的背景信息,而行人特征主要位于轴线区域。其他对象表示形式,如角点对,由于角点远离轴线并且受到背景特征的很大影响,因此不容易通过回归获得。在本文中,我们提出了一种新的检测模式,称为轴线表示和回归(ALR),用于道路场景中的行人检测。具体来说,我们设计了一种用于行人的 3-d 轴线表示,并在网络训练期间将其用作回归目标。还提出了一种线盒转换方法来拟合常用的盒注释。同时,我们探讨了可变形卷积基偏移对检测性能的影响,并提出了一种基偏移初始化策略,以进一步提高 ALR 带来的增益。值得注意的是,所提出的 ALR 模式可以引入基于锚点和无锚点的框架中。我们在 Caltech-USA 和 CityPersons 数据集上验证了 ALR 的有效性。实验结果表明,通过简单的修改,我们的方法在基线显著提高,并且在没有花哨功能的情况下与其他方法具有竞争力的准确性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/371d/8150842/1c89f488b4f1/sensors-21-03312-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/371d/8150842/54807fc893f6/sensors-21-03312-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/371d/8150842/dccdd808feec/sensors-21-03312-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/371d/8150842/bdd8e5e3b047/sensors-21-03312-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/371d/8150842/50909a863298/sensors-21-03312-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/371d/8150842/dd3dec5e43a7/sensors-21-03312-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/371d/8150842/e189b3b92d21/sensors-21-03312-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/371d/8150842/328aa6ca17ab/sensors-21-03312-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/371d/8150842/1c89f488b4f1/sensors-21-03312-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/371d/8150842/54807fc893f6/sensors-21-03312-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/371d/8150842/dccdd808feec/sensors-21-03312-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/371d/8150842/bdd8e5e3b047/sensors-21-03312-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/371d/8150842/50909a863298/sensors-21-03312-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/371d/8150842/dd3dec5e43a7/sensors-21-03312-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/371d/8150842/e189b3b92d21/sensors-21-03312-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/371d/8150842/328aa6ca17ab/sensors-21-03312-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/371d/8150842/1c89f488b4f1/sensors-21-03312-g008.jpg

相似文献

1
Pedestrian Detection by Novel Axis-Line Representation and Regression Pattern.基于新型轴线表示和回归模式的行人检测。
Sensors (Basel). 2021 May 11;21(10):3312. doi: 10.3390/s21103312.
2
Scale-Sensitive Feature Reassembly Network for Pedestrian Detection.尺度敏感特征重组网络用于行人检测。
Sensors (Basel). 2021 Jun 18;21(12):4189. doi: 10.3390/s21124189.
3
Gliding Vertex on the Horizontal Bounding Box for Multi-Oriented Object Detection.用于多方向目标检测的水平边界框上的滑动顶点
IEEE Trans Pattern Anal Mach Intell. 2021 Apr;43(4):1452-1459. doi: 10.1109/TPAMI.2020.2974745. Epub 2021 Mar 5.
4
Too Far to See? Not Really!-Pedestrian Detection With Scale-Aware Localization Policy.太远而无法看清?并非如此!—具有尺度感知本地化策略的行人检测。
IEEE Trans Image Process. 2018 Aug;27(8):3703-3715. doi: 10.1109/TIP.2018.2818018.
5
Mask-Guided Attention Network and Occlusion-Sensitive Hard Example Mining for Occluded Pedestrian Detection.基于掩模引导注意力网络和遮挡敏感硬例挖掘的遮挡行人检测。
IEEE Trans Image Process. 2021;30:3872-3884. doi: 10.1109/TIP.2020.3040854. Epub 2021 Mar 26.
6
Taking a Look at Small-Scale Pedestrians and Occluded Pedestrians.审视小规模行人与被遮挡行人。
IEEE Trans Image Process. 2019 Dec 11. doi: 10.1109/TIP.2019.2957927.
7
Coupled Network for Robust Pedestrian Detection With Gated Multi-Layer Feature Extraction and Deformable Occlusion Handling.具有门控多层特征提取和可变形遮挡处理的稳健行人检测的耦合网络。
IEEE Trans Image Process. 2021;30:754-766. doi: 10.1109/TIP.2020.3038371. Epub 2020 Dec 4.
8
Improving Multiple Pedestrian Tracking in Crowded Scenes with Hierarchical Association.通过分层关联改进拥挤场景中的多行人跟踪
Entropy (Basel). 2023 Feb 19;25(2):380. doi: 10.3390/e25020380.
9
Pedestrian Detection Using Integrated Aggregate Channel Features and Multitask Cascaded Convolutional Neural-Network-Based Face Detectors.基于集成聚合通道特征和多任务级联卷积神经网络的人脸检测器的行人检测。
Sensors (Basel). 2022 May 7;22(9):3568. doi: 10.3390/s22093568.
10
Pedestrian detection algorithm integrating large kernel attention and YOLOV5 lightweight model.集成大核注意力和 YOLOV5 轻量级模型的行人检测算法。
PLoS One. 2023 Nov 29;18(11):e0294865. doi: 10.1371/journal.pone.0294865. eCollection 2023.

引用本文的文献

1
Two-Stage Pedestrian Detection Model Using a New Classification Head for Domain Generalization.基于新型分类头的两阶段行人检测模型用于领域泛化
Sensors (Basel). 2023 Nov 24;23(23):9380. doi: 10.3390/s23239380.

本文引用的文献

1
3D Human Pose Estimation with a Catadioptric Sensor in Unconstrained Environments Using an Annealed Particle Filter.基于退火粒子滤波的非约束环境下折反射传感器的 3D 人体姿态估计
Sensors (Basel). 2020 Dec 7;20(23):6985. doi: 10.3390/s20236985.
2
A Binarized Segmented ResNet Based on Edge Computing for Re-Identification.基于边缘计算的二值化分段 ResNet 再识别。
Sensors (Basel). 2020 Dec 3;20(23):6902. doi: 10.3390/s20236902.
3
Focus on the Visible Regions: Semantic-Guided Alignment Model for Occluded Person Re-Identification.
关注可见区域:遮挡行人再识别的语义引导对齐模型。
Sensors (Basel). 2020 Aug 8;20(16):4431. doi: 10.3390/s20164431.
4
A Dynamic Part-Attention Model for Person Re-Identification.一种用于人物再识别的动态部分注意力模型。
Sensors (Basel). 2019 May 5;19(9):2080. doi: 10.3390/s19092080.
5
Learning Complexity-Aware Cascades for Pedestrian Detection.学习复杂度感知级联用于行人检测。
IEEE Trans Pattern Anal Mach Intell. 2020 Sep;42(9):2195-2211. doi: 10.1109/TPAMI.2019.2910514. Epub 2019 Apr 11.
6
Focal Loss for Dense Object Detection.用于密集目标检测的焦散损失
IEEE Trans Pattern Anal Mach Intell. 2020 Feb;42(2):318-327. doi: 10.1109/TPAMI.2018.2858826. Epub 2018 Jul 23.
7
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.更快的 R-CNN:基于区域建议网络的实时目标检测。
IEEE Trans Pattern Anal Mach Intell. 2017 Jun;39(6):1137-1149. doi: 10.1109/TPAMI.2016.2577031. Epub 2016 Jun 6.
8
Pedestrian detection: an evaluation of the state of the art.行人检测:现状评估。
IEEE Trans Pattern Anal Mach Intell. 2012 Apr;34(4):743-61. doi: 10.1109/TPAMI.2011.155.
9
Object detection with discriminatively trained part-based models.基于判别式训练的部件模型的目标检测。
IEEE Trans Pattern Anal Mach Intell. 2010 Sep;32(9):1627-45. doi: 10.1109/TPAMI.2009.167.