• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

空间注意束:一种关注被遮挡物体的 3D 目标检测方法。

Spatial Attention Frustum: A 3D Object Detection Method Focusing on Occluded Objects.

机构信息

School of Mechanical Engineering, Beijing Institute of Technology, Beijing 100081, China.

出版信息

Sensors (Basel). 2022 Mar 18;22(6):2366. doi: 10.3390/s22062366.

DOI:10.3390/s22062366
PMID:35336536
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8955271/
Abstract

Achieving the accurate perception of occluded objects for autonomous vehicles is a challenging problem. Human vision can always quickly locate important object regions in complex external scenes, while other regions are only roughly analysed or ignored, defined as the visual attention mechanism. However, the perception system of autonomous vehicles cannot know which part of the point cloud is in the region of interest. Therefore, it is meaningful to explore how to use the visual attention mechanism in the perception system of autonomous driving. In this paper, we propose the model of the spatial attention frustum to solve object occlusion in 3D object detection. The spatial attention frustum can suppress unimportant features and allocate limited neural computing resources to critical parts of the scene, thereby providing greater relevance and easier processing for higher-level perceptual reasoning tasks. To ensure that our method maintains good reasoning ability when faced with occluded objects with only a partial structure, we propose a local feature aggregation module to capture more complex local features of the point cloud. Finally, we discuss the projection constraint relationship between the 3D bounding box and the 2D bounding box and propose a joint anchor box projection loss function, which will help to improve the overall performance of our method. The results of the KITTI dataset show that our proposed method can effectively improve the detection accuracy of occluded objects. Our method achieves 89.46%, 79.91% and 75.53% detection accuracy in the easy, moderate, and hard difficulty levels of the car category, and achieves a 6.97% performance improvement especially in the hard category with a high degree of occlusion. Our one-stage method does not need to rely on another refining stage, comparable to the accuracy of the two-stage method.

摘要

实现自动驾驶车辆对被遮挡物体的准确感知是一个具有挑战性的问题。人类视觉总能快速定位复杂外部场景中的重要目标区域,而其他区域则只是粗略分析或忽略,这被定义为视觉注意机制。然而,自动驾驶车辆的感知系统并不知道点云中的哪一部分在感兴趣的区域内。因此,探索如何在自动驾驶的感知系统中使用视觉注意机制是有意义的。在本文中,我们提出了空间注意视锥体的模型,以解决 3D 目标检测中的物体遮挡问题。空间注意视锥体可以抑制不重要的特征,并将有限的神经计算资源分配给场景的关键部分,从而为更高层次的感知推理任务提供更大的相关性和更简单的处理。为了确保我们的方法在仅具有部分结构的遮挡物体面前保持良好的推理能力,我们提出了局部特征聚合模块,以捕获点云中更复杂的局部特征。最后,我们讨论了 3D 边界框和 2D 边界框之间的投影约束关系,并提出了联合锚框投影损失函数,这将有助于提高我们方法的整体性能。KITTI 数据集的结果表明,我们提出的方法可以有效地提高遮挡物体的检测精度。我们的方法在汽车类别的易、中、难三个难度级别上的检测精度分别达到了 89.46%、79.91%和 75.53%,在遮挡程度较高的困难级别上尤其取得了 6.97%的性能提升。我们的单阶段方法不需要依赖另一个精炼阶段,与两阶段方法的精度相当。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/547dba2e8259/sensors-22-02366-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/f65793fbf7ac/sensors-22-02366-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/51ea88fc1e0d/sensors-22-02366-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/1e11b1c9229c/sensors-22-02366-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/2a011c38ff11/sensors-22-02366-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/d477f8e4d3c2/sensors-22-02366-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/a398d637759b/sensors-22-02366-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/bec1b6f165f3/sensors-22-02366-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/fb9421fa1263/sensors-22-02366-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/547dba2e8259/sensors-22-02366-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/f65793fbf7ac/sensors-22-02366-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/51ea88fc1e0d/sensors-22-02366-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/1e11b1c9229c/sensors-22-02366-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/2a011c38ff11/sensors-22-02366-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/d477f8e4d3c2/sensors-22-02366-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/a398d637759b/sensors-22-02366-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/bec1b6f165f3/sensors-22-02366-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/fb9421fa1263/sensors-22-02366-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2142/8955271/547dba2e8259/sensors-22-02366-g009.jpg

相似文献

1
Spatial Attention Frustum: A 3D Object Detection Method Focusing on Occluded Objects.空间注意束:一种关注被遮挡物体的 3D 目标检测方法。
Sensors (Basel). 2022 Mar 18;22(6):2366. doi: 10.3390/s22062366.
2
AEPF: Attention-Enabled Point Fusion for 3D Object Detection.AEPF:用于3D目标检测的注意力增强点融合
Sensors (Basel). 2024 Sep 9;24(17):5841. doi: 10.3390/s24175841.
3
3D Point Cloud Object Detection Method Based on Multi-Scale Dynamic Sparse Voxelization.基于多尺度动态稀疏体素化的三维点云目标检测方法
Sensors (Basel). 2024 Mar 11;24(6):1804. doi: 10.3390/s24061804.
4
3D Object Detection Based on Attention and Multi-Scale Feature Fusion.基于注意力和多尺度特征融合的三维目标检测。
Sensors (Basel). 2022 May 23;22(10):3935. doi: 10.3390/s22103935.
5
CF-YOLOX: An Autonomous Driving Detection Model for Multi-Scale Object Detection.CF-YOLOX:一种用于多尺度目标检测的自动驾驶检测模型。
Sensors (Basel). 2023 Apr 7;23(8):3794. doi: 10.3390/s23083794.
6
BAFusion: Bidirectional Attention Fusion for 3D Object Detection Based on LiDAR and Camera.BAFusion:基于激光雷达和摄像头的用于3D目标检测的双向注意力融合
Sensors (Basel). 2024 Jul 20;24(14):4718. doi: 10.3390/s24144718.
7
AMFF-Net: An Effective 3D Object Detector Based on Attention and Multi-Scale Feature Fusion.AMFF-Net:一种基于注意力和多尺度特征融合的有效3D目标检测器。
Sensors (Basel). 2023 Nov 22;23(23):9319. doi: 10.3390/s23239319.
8
OBMO: One Bounding Box Multiple Objects for Monocular 3D Object Detection.OBMO:用于单目3D目标检测的一个边界框多个目标
IEEE Trans Image Process. 2023 Nov 21;PP. doi: 10.1109/TIP.2023.3333225.
9
From Points to Parts: 3D Object Detection From Point Cloud With Part-Aware and Part-Aggregation Network.从点到部件:基于部件感知与部件聚合网络的点云三维目标检测
IEEE Trans Pattern Anal Mach Intell. 2021 Aug;43(8):2647-2664. doi: 10.1109/TPAMI.2020.2977026. Epub 2021 Jul 1.
10
Multi-View Fusion-Based 3D Object Detection for Robot Indoor Scene Perception.基于多视图融合的机器人室内场景感知三维目标检测。
Sensors (Basel). 2019 Sep 21;19(19):4092. doi: 10.3390/s19194092.

本文引用的文献

1
VoteNet: A Deep Learning Label Fusion Method for Multi-Atlas Segmentation.VoteNet:一种用于多图谱分割的深度学习标签融合方法。
Med Image Comput Comput Assist Interv. 2019 Oct;11766:202-210. doi: 10.1007/978-3-030-32248-9_23. Epub 2019 Oct 10.
2
From Points to Parts: 3D Object Detection From Point Cloud With Part-Aware and Part-Aggregation Network.从点到部件:基于部件感知与部件聚合网络的点云三维目标检测
IEEE Trans Pattern Anal Mach Intell. 2021 Aug;43(8):2647-2664. doi: 10.1109/TPAMI.2020.2977026. Epub 2021 Jul 1.
3
SECOND: Sparsely Embedded Convolutional Detection.
第二:稀疏嵌入卷积检测。
Sensors (Basel). 2018 Oct 6;18(10):3337. doi: 10.3390/s18103337.