• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于特征图扭曲的深度与相机姿态无监督学习。

Unsupervised Learning of Depth and Camera Pose with Feature Map Warping.

作者信息

Guo Ente, Chen Zhifeng, Zhou Yanlin, Wu Dapeng Oliver

机构信息

College of Physics and Information Engineering, Fuzhou University, Fuzhou 350108, China.

Department of Electrical and Computer Engineering, University of Florida, Gainesville, FL 32611, USA.

出版信息

Sensors (Basel). 2021 Jan 30;21(3):923. doi: 10.3390/s21030923.

DOI:10.3390/s21030923
PMID:33573136
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7866542/
Abstract

Estimating the depth of image and egomotion of agent are important for autonomous and robot in understanding the surrounding environment and avoiding collision. Most existing unsupervised methods estimate depth and camera egomotion by minimizing photometric error between adjacent frames. However, the photometric consistency sometimes does not meet the real situation, such as brightness change, moving objects and occlusion. To reduce the influence of brightness change, we propose a feature pyramid matching loss (FPML) which captures the trainable feature error between a current and the adjacent frames and therefore it is more robust than photometric error. In addition, we propose the occlusion-aware mask (OAM) network which can indicate occlusion according to change of masks to improve estimation accuracy of depth and camera pose. The experimental results verify that the proposed unsupervised approach is highly competitive against the state-of-the-art methods, both qualitatively and quantitatively. Specifically, our method reduces absolute relative error (Abs Rel) by 0.017-0.088.

摘要

估计图像深度和智能体的自我运动对于自主系统和机器人理解周围环境以及避免碰撞至关重要。大多数现有的无监督方法通过最小化相邻帧之间的光度误差来估计深度和相机自我运动。然而,光度一致性有时不符合实际情况,例如亮度变化、移动物体和遮挡。为了减少亮度变化的影响,我们提出了一种特征金字塔匹配损失(FPML),它捕捉当前帧和相邻帧之间的可训练特征误差,因此比光度误差更稳健。此外,我们提出了遮挡感知掩码(OAM)网络,它可以根据掩码的变化指示遮挡,以提高深度和相机姿态的估计精度。实验结果验证了所提出的无监督方法在定性和定量方面都与现有最先进方法具有高度竞争力。具体而言,我们的方法将绝对相对误差(Abs Rel)降低了0.017 - 0.088。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d341/7866542/164ae3fb4b7c/sensors-21-00923-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d341/7866542/cd47931ad07d/sensors-21-00923-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d341/7866542/282a245cfe9f/sensors-21-00923-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d341/7866542/99348dc8e4fe/sensors-21-00923-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d341/7866542/fb0ab62884c9/sensors-21-00923-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d341/7866542/164ae3fb4b7c/sensors-21-00923-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d341/7866542/cd47931ad07d/sensors-21-00923-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d341/7866542/282a245cfe9f/sensors-21-00923-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d341/7866542/99348dc8e4fe/sensors-21-00923-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d341/7866542/fb0ab62884c9/sensors-21-00923-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d341/7866542/164ae3fb4b7c/sensors-21-00923-g005.jpg

相似文献

1
Unsupervised Learning of Depth and Camera Pose with Feature Map Warping.基于特征图扭曲的深度与相机姿态无监督学习。
Sensors (Basel). 2021 Jan 30;21(3):923. doi: 10.3390/s21030923.
2
Unsupervised Monocular Depth and Camera Pose Estimation with Multiple Masks and Geometric Consistency Constraints.基于多掩码和几何一致性约束的无监督单目深度与相机位姿估计
Sensors (Basel). 2023 Jun 4;23(11):5329. doi: 10.3390/s23115329.
3
6-DOF Pose Estimation of a Robotic Navigation Aid by Tracking Visual and Geometric Features.通过跟踪视觉和几何特征实现机器人导航辅助设备的六自由度姿态估计
IEEE Trans Autom Sci Eng. 2015 Oct;12(4):1169-1180. doi: 10.1109/TASE.2015.2469726. Epub 2015 Oct 5.
4
Monocular Depth Estimation Using a Laplacian Image Pyramid with Local Planar Guidance Layers.基于拉普拉斯图像金字塔和局部平面引导层的单目深度估计
Sensors (Basel). 2023 Jan 11;23(2):845. doi: 10.3390/s23020845.
5
Unsupervised Monocular Depth Estimation With Channel and Spatial Attention.基于通道和空间注意力的无监督单目深度估计
IEEE Trans Neural Netw Learn Syst. 2024 Jun;35(6):7860-7870. doi: 10.1109/TNNLS.2022.3221416. Epub 2024 Jun 3.
6
Beyond Photometric Consistency: Geometry-Based Occlusion-Aware Unsupervised Light Field Disparity Estimation.超越光度一致性:基于几何的遮挡感知无监督光场视差估计。
IEEE Trans Neural Netw Learn Syst. 2024 Nov;35(11):15660-15674. doi: 10.1109/TNNLS.2023.3289056. Epub 2024 Oct 29.
7
Monocular Depth Estimation via Self-Supervised Self-Distillation.通过自监督自蒸馏进行单目深度估计
Sensors (Basel). 2024 Jun 24;24(13):4090. doi: 10.3390/s24134090.
8
EndoSLAM dataset and an unsupervised monocular visual odometry and depth estimation approach for endoscopic videos.内镜 SLAM 数据集和一种用于内镜视频的无监督单目视觉里程计和深度估计方法。
Med Image Anal. 2021 Jul;71:102058. doi: 10.1016/j.media.2021.102058. Epub 2021 Apr 15.
9
Self-Supervised Monocular Depth Estimation With Self-Perceptual Anomaly Handling.具有自感知异常处理的自监督单目深度估计
IEEE Trans Neural Netw Learn Syst. 2024 Dec;35(12):17292-17306. doi: 10.1109/TNNLS.2023.3301711. Epub 2024 Dec 2.
10
Joint Unsupervised Learning of Depth, Pose, Ground Normal Vector and Ground Segmentation by a Monocular Camera Sensor.基于单目相机传感器的深度、姿势、地面法向量和地面分割的联合无监督学习。
Sensors (Basel). 2020 Jul 3;20(13):3737. doi: 10.3390/s20133737.

本文引用的文献

1
Every Pixel Counts ++: Joint Learning of Geometry and Motion with 3D Holistic Understanding.每一个像素都很重要++:通过3D整体理解进行几何与运动的联合学习。
IEEE Trans Pattern Anal Mach Intell. 2019 Jul 23. doi: 10.1109/TPAMI.2019.2930258.
2
Learning Depth from Single Monocular Images Using Deep Convolutional Neural Fields.利用深度卷积神经场从单目图像中学习深度。
IEEE Trans Pattern Anal Mach Intell. 2016 Oct;38(10):2024-39. doi: 10.1109/TPAMI.2015.2505283. Epub 2015 Dec 3.
3
Image quality assessment: from error visibility to structural similarity.
图像质量评估:从误差可见性到结构相似性。
IEEE Trans Image Process. 2004 Apr;13(4):600-12. doi: 10.1109/tip.2003.819861.