• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

MonoGRNet:单目3D目标检测的通用框架

MonoGRNet: A General Framework for Monocular 3D Object Detection.

作者信息

Qin Zengyi, Wang Jinglu, Lu Yan

出版信息

IEEE Trans Pattern Anal Mach Intell. 2022 Sep;44(9):5170-5184. doi: 10.1109/TPAMI.2021.3074363. Epub 2022 Aug 4.

DOI:10.1109/TPAMI.2021.3074363
PMID:33877968
Abstract

Detecting and localizing objects in the real 3D space, which plays a crucial role in scene understanding, is particularly challenging given only a monocular image due to the geometric information loss during imagery projection. We propose MonoGRNet for the amodal 3D object detection from a monocular image via geometric reasoning in both the observed 2D projection and the unobserved depth dimension. MonoGRNet decomposes the monocular 3D object detection task into four sub-tasks including 2D object detection, instance-level depth estimation, projected 3D center estimation and local corner regression. The task decomposition significantly facilitates the monocular 3D object detection, allowing the target 3D bounding boxes to be efficiently predicted in a single forward pass, without using object proposals, post-processing or the computationally expensive pixel-level depth estimation utilized by previous methods. In addition, MonoGRNet flexibly adapts to both fully and weakly supervised learning, which improves the feasibility of our framework in diverse settings. Experiments are conducted on KITTI, Cityscapes and MS COCO datasets. Results demonstrate the promising performance of our framework in various scenarios.

摘要

在真实3D空间中检测和定位物体,这在场景理解中起着至关重要的作用。由于图像投影过程中的几何信息丢失,仅给定单目图像时,该任务极具挑战性。我们提出了MonoGRNet,用于通过在观察到的2D投影和未观察到的深度维度中进行几何推理,从单目图像中进行无模态3D物体检测。MonoGRNet将单目3D物体检测任务分解为四个子任务,包括2D物体检测、实例级深度估计、投影3D中心估计和局部角点回归。任务分解显著促进了单目3D物体检测,使得目标3D边界框能够在单次前向传播中高效预测,无需使用物体提议、后处理或先前方法中计算成本高昂的像素级深度估计。此外,MonoGRNet灵活适应全监督和弱监督学习,这提高了我们的框架在不同设置下的可行性。在KITTI、Cityscapes和MS COCO数据集上进行了实验。结果证明了我们的框架在各种场景下具有良好的性能。

相似文献

1
MonoGRNet: A General Framework for Monocular 3D Object Detection.MonoGRNet:单目3D目标检测的通用框架
IEEE Trans Pattern Anal Mach Intell. 2022 Sep;44(9):5170-5184. doi: 10.1109/TPAMI.2021.3074363. Epub 2022 Aug 4.
2
MDS-Net: Multi-Scale Depth Stratification 3D Object Detection from Monocular Images.MDS-Net:基于单目图像的多尺度深度分层 3D 目标检测
Sensors (Basel). 2022 Aug 18;22(16):6197. doi: 10.3390/s22166197.
3
MonoAux: Fully Exploiting Auxiliary Information and Uncertainty for Monocular 3D Object Detection.单目辅助(MonoAux):充分利用辅助信息和不确定性进行单目3D目标检测
Cyborg Bionic Syst. 2024 Mar 27;5:0097. doi: 10.34133/cbsystems.0097. eCollection 2024.
4
OBMO: One Bounding Box Multiple Objects for Monocular 3D Object Detection.OBMO:用于单目3D目标检测的一个边界框多个目标
IEEE Trans Image Process. 2023 Nov 21;PP. doi: 10.1109/TIP.2023.3333225.
5
MonoFENet: Monocular 3D Object Detection with Feature Enhancement Networks.单目特征增强网络的单目3D目标检测(MonoFENet)
IEEE Trans Image Process. 2019 Nov 13. doi: 10.1109/TIP.2019.2952201.
6
Vertex points are not enough: Monocular 3D object detection via intra- and inter-plane constraints.顶点不够:通过平面内和平面间约束进行单目 3D 目标检测。
Neural Netw. 2023 May;162:350-358. doi: 10.1016/j.neunet.2023.02.038. Epub 2023 Mar 2.
7
Monocular Quasi-Dense 3D Object Tracking.单目准密集三维物体跟踪
IEEE Trans Pattern Anal Mach Intell. 2023 Feb;45(2):1992-2008. doi: 10.1109/TPAMI.2022.3168781. Epub 2023 Jan 6.
8
Deep Learning-Based Monocular 3D Object Detection with Refinement of Depth Information.基于深度学习的具有深度信息细化的单目 3D 目标检测。
Sensors (Basel). 2022 Mar 28;22(7):2576. doi: 10.3390/s22072576.
9
Weakly Supervised Monocular 3D Object Detection by Spatial-Temporal View Consistency.基于时空视图一致性的弱监督单目3D目标检测
IEEE Trans Pattern Anal Mach Intell. 2025 Jan;47(1):84-98. doi: 10.1109/TPAMI.2024.3466915. Epub 2024 Dec 4.
10
eGAC3D: enhancing depth adaptive convolution and depth estimation for monocular 3D object pose detection.eGAC3D:用于单目3D目标姿态检测的增强深度自适应卷积和深度估计
PeerJ Comput Sci. 2022 Nov 3;8:e1144. doi: 10.7717/peerj-cs.1144. eCollection 2022.

引用本文的文献

1
FCNet: Stereo 3D Object Detection with Feature Correlation Networks.FCNet:基于特征关联网络的立体三维目标检测
Entropy (Basel). 2022 Aug 14;24(8):1121. doi: 10.3390/e24081121.