• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

FCNet:基于特征关联网络的立体三维目标检测

FCNet: Stereo 3D Object Detection with Feature Correlation Networks.

作者信息

Wu Yingyu, Liu Ziyan, Chen Yunlei, Zheng Xuhui, Zhang Qian, Yang Mo, Tang Guangming

机构信息

College of Big Data and Information Engineering, Guizhou University, Guiyang 550025, China.

State Key Laboratory of Public Big Data, Guizhou University, Guiyang 550025, China.

出版信息

Entropy (Basel). 2022 Aug 14;24(8):1121. doi: 10.3390/e24081121.

DOI:10.3390/e24081121
PMID:36010784
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9407267/
Abstract

Deep-learning techniques have significantly improved object detection performance, especially with binocular images in 3D scenarios. To supervise the depth information in stereo 3D object detection, reconstructing the 3D dense depth of LiDAR point clouds causes higher computational costs and lower inference speed. After exploring the intrinsic relationship between the implicit depth information and semantic texture features of the binocular images, we propose an efficient and accurate 3D object detection algorithm, FCNet, in stereo images. First, we construct a multi-scale cost-volume containing implicit depth information using the normalized dot-product by generating multi-scale feature maps from the input stereo images. Secondly, the variant attention model enhances its global and local description, and the sparse region monitors the depth loss deep regression. Thirdly, for balancing the channel information preservation of the re-fused left-right feature maps and computational burden, a reweighting strategy is employed to enhance the feature correlation in merging the last-layer features of binocular images. Extensive experiment results on the challenging KITTI benchmark demonstrate that the proposed algorithm achieves better performance, including a lower computational cost and higher inference speed in 3D object detection.

摘要

深度学习技术显著提高了目标检测性能,尤其是在3D场景中的双目图像方面。为了在立体3D目标检测中监督深度信息,重建激光雷达点云的3D密集深度会导致更高的计算成本和更低的推理速度。在探索了双目图像的隐式深度信息与语义纹理特征之间的内在关系之后,我们提出了一种高效且准确的立体图像3D目标检测算法FCNet。首先,通过从输入的立体图像生成多尺度特征图,利用归一化点积构建包含隐式深度信息的多尺度代价体。其次,可变注意力模型增强其全局和局部描述,稀疏区域监测深度损失深度回归。第三,为了平衡重新融合的左右特征图的通道信息保留和计算负担,采用了一种重新加权策略来增强在合并双目图像最后一层特征时的特征相关性。在具有挑战性的KITTI基准上的大量实验结果表明,所提出的算法在3D目标检测中实现了更好的性能,包括更低的计算成本和更高的推理速度。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/170a4368d347/entropy-24-01121-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/20210c1a661a/entropy-24-01121-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/eb77ae81bcd7/entropy-24-01121-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/0ced1fb1958d/entropy-24-01121-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/c1a7bb76aa64/entropy-24-01121-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/a23db12ce41a/entropy-24-01121-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/18c7a8d62423/entropy-24-01121-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/d5a2fbc9fb9e/entropy-24-01121-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/e00840465077/entropy-24-01121-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/42b2b49b7322/entropy-24-01121-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/8202dabe0398/entropy-24-01121-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/170a4368d347/entropy-24-01121-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/20210c1a661a/entropy-24-01121-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/eb77ae81bcd7/entropy-24-01121-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/0ced1fb1958d/entropy-24-01121-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/c1a7bb76aa64/entropy-24-01121-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/a23db12ce41a/entropy-24-01121-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/18c7a8d62423/entropy-24-01121-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/d5a2fbc9fb9e/entropy-24-01121-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/e00840465077/entropy-24-01121-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/42b2b49b7322/entropy-24-01121-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/8202dabe0398/entropy-24-01121-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3429/9407267/170a4368d347/entropy-24-01121-g011.jpg

相似文献

1
FCNet: Stereo 3D Object Detection with Feature Correlation Networks.FCNet:基于特征关联网络的立体三维目标检测
Entropy (Basel). 2022 Aug 14;24(8):1121. doi: 10.3390/e24081121.
2
MS3D: A 3D object detection method using multi-scale semantic feature points to construct 3D feature layer.MS3D:一种使用多尺度语义特征点构建 3D 特征层的 3D 目标检测方法。
Neural Netw. 2024 Nov;179:106623. doi: 10.1016/j.neunet.2024.106623. Epub 2024 Aug 10.
3
PSANet: Pyramid Splitting and Aggregation Network for 3D Object Detection in Point Cloud.PSANet:用于点云中 3D 目标检测的金字塔分裂与聚合网络。
Sensors (Basel). 2020 Dec 28;21(1):136. doi: 10.3390/s21010136.
4
An end-to-end stereo matching algorithm based on improved convolutional neural network.基于改进卷积神经网络的端到端立体匹配算法。
Math Biosci Eng. 2020 Nov 6;17(6):7787-7803. doi: 10.3934/mbe.2020396.
5
Relation Graph Network for 3D Object Detection in Point Clouds.用于点云三维目标检测的关系图网络
IEEE Trans Image Process. 2021;30:92-107. doi: 10.1109/TIP.2020.3031371. Epub 2020 Nov 18.
6
Voxel-FPN: Multi-Scale Voxel Feature Aggregation for 3D Object Detection from LIDAR Point Clouds.体素特征金字塔网络(Voxel-FPN):用于从激光雷达点云进行三维目标检测的多尺度体素特征聚合
Sensors (Basel). 2020 Jan 28;20(3):704. doi: 10.3390/s20030704.
7
A Two-Phase Cross-Modality Fusion Network for Robust 3D Object Detection.用于稳健3D目标检测的两阶段跨模态融合网络
Sensors (Basel). 2020 Oct 23;20(21):6043. doi: 10.3390/s20216043.
8
Efficient Multi-Scale Stereo-Matching Network Using Adaptive Cost Volume Filtering.基于自适应代价体滤波的高效多尺度立体匹配网络
Sensors (Basel). 2022 Jul 23;22(15):5500. doi: 10.3390/s22155500.
9
3D Object Proposals Using Stereo Imagery for Accurate Object Class Detection.基于立体图像的 3D 对象建议,用于准确的目标类别检测。
IEEE Trans Pattern Anal Mach Intell. 2018 May;40(5):1259-1272. doi: 10.1109/TPAMI.2017.2706685. Epub 2017 May 19.
10
Joint stereo 3D object detection and implicit surface reconstruction.联合立体三维目标检测与隐式曲面重建
Sci Rep. 2024 Jun 17;14(1):13893. doi: 10.1038/s41598-024-64677-2.

本文引用的文献

1
Pedestrian Detection with Multi-View Convolution Fusion Algorithm.基于多视图卷积融合算法的行人检测
Entropy (Basel). 2022 Jan 22;24(2):165. doi: 10.3390/e24020165.
2
MonoGRNet: A General Framework for Monocular 3D Object Detection.MonoGRNet:单目3D目标检测的通用框架
IEEE Trans Pattern Anal Mach Intell. 2022 Sep;44(9):5170-5184. doi: 10.1109/TPAMI.2021.3074363. Epub 2022 Aug 4.
3
MonoFENet: Monocular 3D Object Detection with Feature Enhancement Networks.单目特征增强网络的单目3D目标检测(MonoFENet)
IEEE Trans Image Process. 2019 Nov 13. doi: 10.1109/TIP.2019.2952201.
4
SECOND: Sparsely Embedded Convolutional Detection.第二:稀疏嵌入卷积检测。
Sensors (Basel). 2018 Oct 6;18(10):3337. doi: 10.3390/s18103337.
5
Kinematic Control of Redundant Manipulators Using Neural Networks.基于神经网络的冗余机械臂运动控制
IEEE Trans Neural Netw Learn Syst. 2017 Oct;28(10):2243-2254. doi: 10.1109/TNNLS.2016.2574363. Epub 2016 Jun 24.
6
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.更快的 R-CNN:基于区域建议网络的实时目标检测。
IEEE Trans Pattern Anal Mach Intell. 2017 Jun;39(6):1137-1149. doi: 10.1109/TPAMI.2016.2577031. Epub 2016 Jun 6.
7
Distributed Recurrent Neural Networks for Cooperative Control of Manipulators: A Game-Theoretic Perspective.分布式递归神经网络在机械臂协同控制中的应用:一种博弈论的视角。
IEEE Trans Neural Netw Learn Syst. 2017 Feb;28(2):415-426. doi: 10.1109/TNNLS.2016.2516565. Epub 2016 Jan 21.