• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

仅使用合成图像对头部体模中机器人器械的语义分割对 CNN 进行训练时,不同逼真度水平的影响。

The effects of different levels of realism on the training of CNNs with only synthetic images for the semantic segmentation of robotic instruments in a head phantom.

机构信息

Department of Mechanical Engineering, The University of Tokyo, Tokyo, Japan.

出版信息

Int J Comput Assist Radiol Surg. 2020 Aug;15(8):1257-1265. doi: 10.1007/s11548-020-02185-0. Epub 2020 May 22.

DOI:10.1007/s11548-020-02185-0
PMID:32445129
Abstract

PURPOSE

The manual generation of training data for the semantic segmentation of medical images using deep neural networks is a time-consuming and error-prone task. In this paper, we investigate the effect of different levels of realism on the training of deep neural networks for semantic segmentation of robotic instruments. An interactive virtual-reality environment was developed to generate synthetic images for robot-aided endoscopic surgery. In contrast with earlier works, we use physically based rendering for increased realism.

METHODS

Using a virtual reality simulator that replicates our robotic setup, three synthetic image databases with an increasing level of realism were generated: flat, basic, and realistic (using the physically-based rendering). Each of those databases was used to train 20 instances of a UNet-based semantic-segmentation deep-learning model. The networks trained with only synthetic images were evaluated on the segmentation of 160 endoscopic images of a phantom. The networks were compared using the Dwass-Steel-Critchlow-Fligner nonparametric test.

RESULTS

Our results show that the levels of realism increased the mean intersection-over-union (mIoU) of the networks on endoscopic images of a phantom ([Formula: see text]). The median mIoU values were 0.235 for the flat dataset, 0.458 for the basic, and 0.729 for the realistic. All the networks trained with synthetic images outperformed naive classifiers. Moreover, in an ablation study, we show that the mIoU of physically based rendering is superior to texture mapping ([Formula: see text]) of the instrument (0.606), the background (0.685), and the background and instruments combined (0.672).

CONCLUSIONS

Using physical-based rendering to generate synthetic images is an effective approach to improve the training of neural networks for the semantic segmentation of surgical instruments in endoscopic images. Our results show that this strategy can be an essential step in the broad applicability of deep neural networks in semantic segmentation tasks and help bridge the domain gap in machine learning.

摘要

目的

使用深度神经网络对医学图像进行语义分割的训练数据的手动生成是一项耗时且容易出错的任务。在本文中,我们研究了不同逼真度水平对机器人器械语义分割的深度神经网络训练的影响。开发了一个交互式虚拟现实环境,用于生成机器人辅助内窥镜手术的合成图像。与早期的工作相比,我们使用基于物理的渲染来提高逼真度。

方法

使用复制我们机器人设置的虚拟现实模拟器,生成了三个逼真度递增的合成图像数据库:平面、基础和真实(使用基于物理的渲染)。每个数据库都用于训练 20 个基于 UNet 的语义分割深度学习模型的实例。仅使用合成图像训练的网络在 160 个幻影内窥镜图像的分割上进行评估。使用 Dwass-Steel-Critchlow-Fligner 非参数检验比较网络。

结果

我们的结果表明,逼真度水平提高了网络在幻影内窥镜图像上的平均交并比(mIoU)([Formula: see text])。平面数据集的中位数 mIoU 值为 0.235,基础数据集为 0.458,真实数据集为 0.729。所有使用合成图像训练的网络都优于朴素分类器。此外,在一项消融研究中,我们表明基于物理的渲染的 mIoU 优于仪器的纹理映射([Formula: see text])(0.606)、背景(0.685)和背景与仪器的组合(0.672)。

结论

使用基于物理的渲染生成合成图像是一种有效的方法,可以提高深度神经网络在内窥镜图像中对手术器械语义分割的训练效果。我们的结果表明,这种策略可能是深度神经网络在语义分割任务中的广泛应用的关键步骤,并有助于弥合机器学习中的领域差距。

相似文献

1
The effects of different levels of realism on the training of CNNs with only synthetic images for the semantic segmentation of robotic instruments in a head phantom.仅使用合成图像对头部体模中机器人器械的语义分割对 CNN 进行训练时,不同逼真度水平的影响。
Int J Comput Assist Radiol Surg. 2020 Aug;15(8):1257-1265. doi: 10.1007/s11548-020-02185-0. Epub 2020 May 22.
2
Image generation by GAN and style transfer for agar plate image segmentation.基于 GAN 和风格迁移的琼脂平板图像分割的图像生成。
Comput Methods Programs Biomed. 2020 Feb;184:105268. doi: 10.1016/j.cmpb.2019.105268. Epub 2019 Dec 17.
3
A dataset of laryngeal endoscopic images with comparative study on convolution neural network-based semantic segmentation.基于卷积神经网络的语义分割的喉内窥镜图像数据集及对比研究。
Int J Comput Assist Radiol Surg. 2019 Mar;14(3):483-492. doi: 10.1007/s11548-018-01910-0. Epub 2019 Jan 16.
4
A Performance Improvement Strategy for Concrete Damage Detection Using Stacking Ensemble Learning of Multiple Semantic Segmentation Networks.基于多语义分割网络堆叠集成学习的混凝土损伤检测性能提升策略。
Sensors (Basel). 2022 Apr 27;22(9):3341. doi: 10.3390/s22093341.
5
Pulmonary nodule segmentation with CT sample synthesis using adversarial networks.基于对抗网络的 CT 样本合成的肺结节分割。
Med Phys. 2019 Mar;46(3):1218-1229. doi: 10.1002/mp.13349. Epub 2019 Jan 31.
6
Enabling machine learning in X-ray-based procedures via realistic simulation of image formation.通过基于 X 射线的成像形成的逼真模拟来实现机器学习。
Int J Comput Assist Radiol Surg. 2019 Sep;14(9):1517-1528. doi: 10.1007/s11548-019-02011-2. Epub 2019 Jun 11.
7
Multi-Scale Context-Guided Deep Network for Automated Lesion Segmentation With Endoscopy Images of Gastrointestinal Tract.多尺度上下文引导的深度网络用于胃肠道内窥镜图像的自动病变分割。
IEEE J Biomed Health Inform. 2021 Feb;25(2):514-525. doi: 10.1109/JBHI.2020.2997760. Epub 2021 Feb 5.
8
Dual-stage semantic segmentation of endoscopic surgical instruments.内窥镜手术器械的双阶段语义分割
Med Phys. 2024 Dec;51(12):9125-9137. doi: 10.1002/mp.17397. Epub 2024 Sep 10.
9
Improving realism in patient-specific abdominal ultrasound simulation using CycleGANs.利用 CycleGAN 提高腹部超声模拟中的患者特异性真实感。
Int J Comput Assist Radiol Surg. 2020 Feb;15(2):183-192. doi: 10.1007/s11548-019-02046-5. Epub 2019 Aug 7.
10
Semi-supervised learning for automatic segmentation of the knee from MRI with convolutional neural networks.基于卷积神经网络的膝关节 MRI 半自动分割的半监督学习。
Comput Methods Programs Biomed. 2020 Jun;189:105328. doi: 10.1016/j.cmpb.2020.105328. Epub 2020 Jan 11.