• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于腹腔镜图像的基于3D位移模块的自监督单目深度估计

Self-supervised Monocular Depth Estimation with 3D Displacement Module for Laparoscopic Images.

作者信息

Xu Chi, Huang Baoru, Elson Daniel S

机构信息

The Hamlyn Centre for Robotic Surgery, Department of Surgery and Cancer, Imperial College London, London SW7 2AZ, UK.

出版信息

IEEE Trans Med Robot Bionics. 2022 May;4(2):331-334. doi: 10.1109/TMRB.2022.3170206.

DOI:10.1109/TMRB.2022.3170206
PMID:36148138
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7613618/
Abstract

We present a novel self-supervised training framework with 3D displacement (3DD) module for accurately estimating per-pixel depth maps from single laparoscopic images. Recently, several self-supervised learning based monocular depth estimation models have achieved good results on the KITTI dataset, under the hypothesis that the camera is dynamic and the objects are stationary, however this hypothesis is often reversed in the surgical setting (laparoscope is stationary, the surgical instruments and tissues are dynamic). Therefore, a 3DD module is proposed to establish the relation between frames instead of ego-motion estimation. In the 3DD module, a convolutional neural network (CNN) analyses source and target frames to predict the 3D displacement of a 3D point cloud from a target frame to a source frame in the coordinates of the camera. Since it is difficult to constrain the depth displacement from two 2D images, a novel depth consistency module is proposed to maintain depth consistency between displacement-updated depth and model-estimated depth to constrain 3D displacement effectively. Our proposed method achieves remarkable performance for monocular depth estimation on the Hamlyn surgical dataset and acquired ground truth depth maps, outperforming monodepth, monodepth2 and packnet models.

摘要

我们提出了一种带有3D位移(3DD)模块的新型自监督训练框架,用于从单张腹腔镜图像中准确估计逐像素深度图。最近,一些基于自监督学习的单目深度估计模型在KITTI数据集上取得了良好的效果,其假设是相机是动态的而物体是静止的,然而在手术场景中(腹腔镜是静止的,手术器械和组织是动态的),这个假设往往是相反的。因此,提出了一个3DD模块来建立帧之间的关系,而不是进行自我运动估计。在3DD模块中,一个卷积神经网络(CNN)分析源帧和目标帧,以预测相机坐标中从目标帧到源帧的3D点云的3D位移。由于很难从两个2D图像中约束深度位移,因此提出了一个新颖的深度一致性模块,以保持位移更新深度和模型估计深度之间的深度一致性,从而有效地约束3D位移。我们提出的方法在哈姆林手术数据集和获取的地面真值深度图上进行单目深度估计时取得了显著的性能,优于单目深度(Monodepth)、单目深度2(Monodepth2)和PackNet模型。

相似文献

1
Self-supervised Monocular Depth Estimation with 3D Displacement Module for Laparoscopic Images.用于腹腔镜图像的基于3D位移模块的自监督单目深度估计
IEEE Trans Med Robot Bionics. 2022 May;4(2):331-334. doi: 10.1109/TMRB.2022.3170206.
2
WS-SfMLearner: self-supervised monocular depth and ego-motion estimation on surgical videos with unknown camera parameters.WS-SfMLearner:在相机参数未知的手术视频上进行自监督单目深度和自我运动估计。
J Med Imaging (Bellingham). 2025 Mar;12(2):025003. doi: 10.1117/1.JMI.12.2.025003. Epub 2025 Apr 30.
3
Self-Supervised monocular depth and ego-Motion estimation in endoscopy: Appearance flow to the rescue.内窥镜中单目深度和自我运动估计的自监督学习:外观流来救援。
Med Image Anal. 2022 Apr;77:102338. doi: 10.1016/j.media.2021.102338. Epub 2021 Dec 25.
4
Enhanced self-supervised monocular depth estimation with self-attention and joint depth-pose loss for laparoscopic images.基于自注意力机制和联合深度-姿态损失的腹腔镜图像增强自监督单目深度估计
Int J Comput Assist Radiol Surg. 2025 Apr;20(4):775-785. doi: 10.1007/s11548-025-03332-1. Epub 2025 Feb 28.
5
Self-supervised recurrent depth estimation with attention mechanisms.基于注意力机制的自监督循环深度估计
PeerJ Comput Sci. 2022 Jan 31;8:e865. doi: 10.7717/peerj-cs.865. eCollection 2022.
6
Monocular Depth Estimation via Self-Supervised Self-Distillation.通过自监督自蒸馏进行单目深度估计
Sensors (Basel). 2024 Jun 24;24(13):4090. doi: 10.3390/s24134090.
7
Confidence-aware self-supervised learning for dense monocular depth estimation in dynamic laparoscopic scene.基于置信感知的自监督学习的动态腹腔镜场景稠密单目深度估计
Sci Rep. 2023 Sep 16;13(1):15380. doi: 10.1038/s41598-023-42713-x.
8
Monocular Depth Estimation with Self-Supervised Learning for Vineyard Unmanned Agricultural Vehicle.基于自监督学习的葡萄园无人农业车单目深度估计
Sensors (Basel). 2022 Jan 18;22(3):721. doi: 10.3390/s22030721.
9
SENSE: Self-Evolving Learning for Self-Supervised Monocular Depth Estimation.SENSE:用于自监督单目深度估计的自进化学习
IEEE Trans Image Process. 2024;33:439-450. doi: 10.1109/TIP.2023.3338053. Epub 2023 Dec 29.
10
PMIndoor: Pose Rectified Network and Multiple Loss Functions for Self-Supervised Monocular Indoor Depth Estimation.PMIndoor:用于自监督单目室内深度估计的姿态校正网络和多重损失函数
Sensors (Basel). 2023 Oct 30;23(21):8821. doi: 10.3390/s23218821.

引用本文的文献

1
Comment on: "3D Da Vinci robotic surgery: is it a risk to the surgeon's eye health?" by Molle, et al.关于莫勒等人的《3D达芬奇机器人手术:对外科医生眼部健康有风险吗?》的评论
J Robot Surg. 2025 Feb 7;19(1):63. doi: 10.1007/s11701-025-02226-3.
2
Application of Machine Vision Techniques in Low-Cost Devices to Improve Efficiency in Precision Farming.机器视觉技术在低成本设备中的应用以提高精准农业效率。
Sensors (Basel). 2024 Jan 31;24(3):937. doi: 10.3390/s24030937.

本文引用的文献

1
Dense Depth Estimation in Monocular Endoscopy With Self-Supervised Learning Methods.基于自监督学习方法的单目内窥镜下密集深度估计。
IEEE Trans Med Imaging. 2020 May;39(5):1438-1447. doi: 10.1109/TMI.2019.2950936. Epub 2019 Nov 1.
2
Evaluation and Stability Analysis of Video-Based Navigation System for Functional Endoscopic Sinus Surgery on In Vivo Clinical Data.基于活体临床数据的功能性内窥镜鼻窦手术视频导航系统的评估与稳定性分析。
IEEE Trans Med Imaging. 2018 Oct;37(10):2185-2195. doi: 10.1109/TMI.2018.2833868. Epub 2018 May 7.
3
Minimally invasive radioguided parathyroid surgery: A literature review.
微创放射性引导甲状旁腺手术:文献综述。
Int J Surg. 2016 Apr;28 Suppl 1:S84-93. doi: 10.1016/j.ijsu.2015.12.037. Epub 2015 Dec 23.
4
Haptics in minimally invasive surgery--a review.微创手术中的触觉——综述
Minim Invasive Ther Allied Technol. 2008;17(1):3-16. doi: 10.1080/13645700701820242.
5
Image quality assessment: from error visibility to structural similarity.图像质量评估:从误差可见性到结构相似性。
IEEE Trans Image Process. 2004 Apr;13(4):600-12. doi: 10.1109/tip.2003.819861.