• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

无需人工标注的手术工具分割图像合成

Image Compositing for Segmentation of Surgical Tools Without Manual Annotations.

作者信息

Garcia-Peraza-Herrera Luis C, Fidon Lucas, D'Ettorre Claudia, Stoyanov Danail, Vercauteren Tom, Ourselin Sebastien

出版信息

IEEE Trans Med Imaging. 2021 May;40(5):1450-1460. doi: 10.1109/TMI.2021.3057884. Epub 2021 Apr 30.

DOI:10.1109/TMI.2021.3057884
PMID:33556005
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8092331/
Abstract

Producing manual, pixel-accurate, image segmentation labels is tedious and time-consuming. This is often a rate-limiting factor when large amounts of labeled images are required, such as for training deep convolutional networks for instrument-background segmentation in surgical scenes. No large datasets comparable to industry standards in the computer vision community are available for this task. To circumvent this problem, we propose to automate the creation of a realistic training dataset by exploiting techniques stemming from special effects and harnessing them to target training performance rather than visual appeal. Foreground data is captured by placing sample surgical instruments over a chroma key (a.k.a. green screen) in a controlled environment, thereby making extraction of the relevant image segment straightforward. Multiple lighting conditions and viewpoints can be captured and introduced in the simulation by moving the instruments and camera and modulating the light source. Background data is captured by collecting videos that do not contain instruments. In the absence of pre-existing instrument-free background videos, minimal labeling effort is required, just to select frames that do not contain surgical instruments from videos of surgical interventions freely available online. We compare different methods to blend instruments over tissue and propose a novel data augmentation approach that takes advantage of the plurality of options. We show that by training a vanilla U-Net on semi-synthetic data only and applying a simple post-processing, we are able to match the results of the same network trained on a publicly available manually labeled real dataset.

摘要

生成手动的、像素精确的图像分割标签既繁琐又耗时。当需要大量带标签的图像时,比如用于训练手术场景中仪器 - 背景分割的深度卷积网络时,这往往是一个限速因素。目前没有与计算机视觉社区行业标准相当的大型数据集可用于此任务。为了规避这个问题,我们建议通过利用源自特效的技术并将其用于提升训练性能而非视觉效果,来自动创建一个逼真的训练数据集。前景数据是在受控环境中通过将样本手术器械放置在色度键(又称绿幕)上进行采集的,从而使相关图像片段的提取变得简单直接。通过移动器械和相机以及调节光源,可以在模拟中捕捉和引入多种光照条件和视角。背景数据是通过收集不含器械的视频来获取的。在没有预先存在的无器械背景视频的情况下,只需要进行最少的标注工作,即从网上免费获取的手术干预视频中选择不包含手术器械的帧即可。我们比较了将器械融合到组织上的不同方法,并提出了一种利用多种选项的新型数据增强方法。我们表明,仅在半合成数据上训练一个普通的U-Net并应用简单的后处理,我们就能与在公开可用的手动标注真实数据集上训练的同一网络的结果相匹配。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae95/8092331/54730a50cea9/garci4-3057884.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae95/8092331/a5af914aa032/garci1-3057884.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae95/8092331/210e636ca952/garci2-3057884.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae95/8092331/6236557d68d5/garci3-3057884.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae95/8092331/54730a50cea9/garci4-3057884.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae95/8092331/a5af914aa032/garci1-3057884.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae95/8092331/210e636ca952/garci2-3057884.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae95/8092331/6236557d68d5/garci3-3057884.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae95/8092331/54730a50cea9/garci4-3057884.jpg

相似文献

1
Image Compositing for Segmentation of Surgical Tools Without Manual Annotations.无需人工标注的手术工具分割图像合成
IEEE Trans Med Imaging. 2021 May;40(5):1450-1460. doi: 10.1109/TMI.2021.3057884. Epub 2021 Apr 30.
2
Catheter segmentation in X-ray fluoroscopy using synthetic data and transfer learning with light U-nets.基于合成数据和轻量级 U 型网络的迁移学习在 X 射线透视下的导管分割
Comput Methods Programs Biomed. 2020 Aug;192:105420. doi: 10.1016/j.cmpb.2020.105420. Epub 2020 Feb 29.
3
The effects of different levels of realism on the training of CNNs with only synthetic images for the semantic segmentation of robotic instruments in a head phantom.仅使用合成图像对头部体模中机器人器械的语义分割对 CNN 进行训练时,不同逼真度水平的影响。
Int J Comput Assist Radiol Surg. 2020 Aug;15(8):1257-1265. doi: 10.1007/s11548-020-02185-0. Epub 2020 May 22.
4
Detection, segmentation, and 3D pose estimation of surgical tools using convolutional neural networks and algebraic geometry.使用卷积神经网络和代数几何进行手术工具的检测、分割和三维姿态估计。
Med Image Anal. 2021 May;70:101994. doi: 10.1016/j.media.2021.101994. Epub 2021 Feb 7.
5
Image generation by GAN and style transfer for agar plate image segmentation.基于 GAN 和风格迁移的琼脂平板图像分割的图像生成。
Comput Methods Programs Biomed. 2020 Feb;184:105268. doi: 10.1016/j.cmpb.2019.105268. Epub 2019 Dec 17.
6
A convolutional neural network for segmentation of yeast cells without manual training annotations.一种无需手动训练注释的用于酵母细胞分割的卷积神经网络。
Bioinformatics. 2022 Feb 7;38(5):1427-1433. doi: 10.1093/bioinformatics/btab835.
7
Unpaired deep adversarial learning for multi-class segmentation of instruments in robot-assisted surgical videos.无配对深度对抗学习在机器人辅助手术视频中对器械进行多类分割。
Int J Med Robot. 2023 Aug;19(4):e2514. doi: 10.1002/rcs.2514. Epub 2023 Mar 28.
8
Generating Synthetic Labeled Data From Existing Anatomical Models: An Example With Echocardiography Segmentation.从现有解剖模型生成合成标记数据:以心脏超声分割为例。
IEEE Trans Med Imaging. 2021 Oct;40(10):2783-2794. doi: 10.1109/TMI.2021.3051806. Epub 2021 Sep 30.
9
Fully automatic multi-organ segmentation for head and neck cancer radiotherapy using shape representation model constrained fully convolutional neural networks.使用基于形状表示模型约束的全卷积神经网络进行头颈部癌症放疗的全自动多器官分割。
Med Phys. 2018 Oct;45(10):4558-4567. doi: 10.1002/mp.13147. Epub 2018 Sep 19.
10
Data augmentation using generative adversarial networks (CycleGAN) to improve generalizability in CT segmentation tasks.使用生成对抗网络(CycleGAN)进行数据增强以提高 CT 分割任务的泛化能力。
Sci Rep. 2019 Nov 15;9(1):16884. doi: 10.1038/s41598-019-52737-x.

引用本文的文献

1
SegMatch: semi-supervised surgical instrument segmentation.SegMatch:半监督手术器械分割
Sci Rep. 2025 Apr 23;15(1):14042. doi: 10.1038/s41598-025-94568-z.
2
Optimizing intraoperative AI: evaluation of YOLOv8 for real-time recognition of robotic and laparoscopic instruments.优化术中人工智能:评估YOLOv8对机器人和腹腔镜器械的实时识别能力。
J Robot Surg. 2025 Mar 31;19(1):131. doi: 10.1007/s11701-025-02284-7.
3
Neural Radiance Fields for High-Fidelity Soft Tissue Reconstruction in Endoscopy.用于内窥镜检查中高保真软组织重建的神经辐射场

本文引用的文献

1
Enabling machine learning in X-ray-based procedures via realistic simulation of image formation.通过基于 X 射线的成像形成的逼真模拟来实现机器学习。
Int J Comput Assist Radiol Surg. 2019 Sep;14(9):1517-1528. doi: 10.1007/s11548-019-02011-2. Epub 2019 Jun 11.
2
Surgical data science for next-generation interventions.面向下一代干预措施的外科数据科学。
Nat Biomed Eng. 2017 Sep;1(9):691-696. doi: 10.1038/s41551-017-0132-7.
3
Weakly supervised convolutional LSTM approach for tool tracking in laparoscopic videos.基于弱监督卷积 LSTM 的腹腔镜视频中工具跟踪方法。
Sensors (Basel). 2025 Jan 19;25(2):565. doi: 10.3390/s25020565.
4
Deep homography estimation in dynamic surgical scenes for laparoscopic camera motion extraction.用于腹腔镜相机运动提取的动态手术场景中的深度单应性估计。
Comput Methods Biomech Biomed Eng Imaging Vis. 2022 Feb 23;10(3):321-329. doi: 10.1080/21681163.2021.2002195. eCollection 2022.
5
Eliminating the need for manual segmentation to determine size and volume from MRI. A proof of concept on segmenting the lateral ventricles.消除手动分割以从 MRI 确定大小和体积的需求。对侧脑室分割的概念验证。
PLoS One. 2023 May 11;18(5):e0285414. doi: 10.1371/journal.pone.0285414. eCollection 2023.
6
Min-Max Similarity: A Contrastive Semi-Supervised Deep Learning Network for Surgical Tools Segmentation.最小-最大相似度:一种用于手术工具分割的对比半监督深度学习网络。
IEEE Trans Med Imaging. 2023 Oct;42(10):2832-2841. doi: 10.1109/TMI.2023.3266137. Epub 2023 Oct 2.
7
Improving needle visibility in LED-based photoacoustic imaging using deep learning with semi-synthetic datasets.利用深度学习和半合成数据集提高基于LED的光声成像中的针可见性。
Photoacoustics. 2022 Apr 7;26:100351. doi: 10.1016/j.pacs.2022.100351. eCollection 2022 Jun.
8
Robotic Endoscope Control Via Autonomous Instrument Tracking.通过自主器械跟踪实现机器人内窥镜控制。
Front Robot AI. 2022 Apr 11;9:832208. doi: 10.3389/frobt.2022.832208. eCollection 2022.
Int J Comput Assist Radiol Surg. 2019 Jun;14(6):1059-1067. doi: 10.1007/s11548-019-01958-6. Epub 2019 Apr 9.
4
Exploiting the potential of unlabeled endoscopic video data with self-supervised learning.利用自监督学习挖掘未标记内镜视频数据的潜力。
Int J Comput Assist Radiol Surg. 2018 Jun;13(6):925-933. doi: 10.1007/s11548-018-1772-0. Epub 2018 Apr 27.
5
Real-time ultrasound transducer localization in fluoroscopy images by transfer learning from synthetic training data.基于合成训练数据的迁移学习实现透视图像中实时超声换能器定位。
Med Image Anal. 2014 Dec;18(8):1320-8. doi: 10.1016/j.media.2014.04.007. Epub 2014 May 5.
6
Atlas encoding by randomized forests for efficient label propagation.用于高效标签传播的随机森林地图集编码
Med Image Comput Comput Assist Interv. 2013;16(Pt 3):66-73. doi: 10.1007/978-3-642-40760-4_9.
7
Feature classification for tracking articulated surgical tools.用于跟踪关节式手术工具的特征分类
Med Image Comput Comput Assist Interv. 2012;15(Pt 2):592-600. doi: 10.1007/978-3-642-33418-4_73.
8
Towards image guided robotic surgery: multi-arm tracking through hybrid localization.迈向图像引导机器人手术:通过混合定位进行多臂跟踪。
Int J Comput Assist Radiol Surg. 2009 May;4(3):281-6. doi: 10.1007/s11548-009-0294-1. Epub 2009 Mar 19.