Department of Gynaecological Surgery, CHU Clermont-Ferrand, 1 Place Lucie et Raymond Aubrac, 63000, Clermont-Ferrand, France.
EnCoV, Institut Pascal, CNRS, Université Clermont Auvergne, Clermont-Ferrand, France.
Surg Endosc. 2020 Dec;34(12):5377-5383. doi: 10.1007/s00464-019-07330-8. Epub 2020 Jan 29.
In laparoscopy, the digital camera offers surgeons the opportunity to receive support from image-guided surgery systems. Such systems require image understanding, the ability for a computer to understand what the laparoscope sees. Image understanding has recently progressed owing to the emergence of artificial intelligence and especially deep learning techniques. However, the state of the art of deep learning in gynaecology only offers image-based detection, reporting the presence or absence of an anatomical structure, without finding its location. A solution to the localisation problem is given by the concept of semantic segmentation, giving the detection and pixel-level location of a structure in an image. The state-of-the-art results in semantic segmentation are achieved by deep learning, whose usage requires a massive amount of annotated data. We propose the first dataset dedicated to this task and the first evaluation of deep learning-based semantic segmentation in gynaecology.
We used the deep learning method called Mask R-CNN. Our dataset has 461 laparoscopic images manually annotated with three classes: uterus, ovaries and surgical tools. We split our dataset in 361 images to train Mask R-CNN and 100 images to evaluate its performance.
The segmentation accuracy is reported in terms of percentage of overlap between the segmented regions from Mask R-CNN and the manually annotated ones. The accuracy is 84.5%, 29.6% and 54.5% for uterus, ovaries and surgical tools, respectively. An automatic detection of these structures was then inferred from the semantic segmentation results which led to state-of-the-art detection performance, except for the ovaries. Specifically, the detection accuracy is 97%, 24% and 86% for uterus, ovaries and surgical tools, respectively.
Our preliminary results are very promising, given the relatively small size of our initial dataset. The creation of an international surgical database seems essential.
在腹腔镜手术中,数字摄像机为外科医生提供了从图像引导手术系统获得支持的机会。此类系统需要图像理解,即计算机理解腹腔镜所看到的内容的能力。由于人工智能尤其是深度学习技术的出现,图像理解最近取得了进展。然而,妇科深度学习的最新技术仅提供基于图像的检测,报告解剖结构的存在或不存在,而无法找到其位置。一种解决定位问题的方法是语义分割的概念,它提供了图像中结构的检测和像素级位置。语义分割的最新技术成果是通过深度学习实现的,其使用需要大量注释数据。我们提出了第一个专门用于该任务的数据集,并首次评估了妇科深度学习的语义分割。
我们使用了名为 Mask R-CNN 的深度学习方法。我们的数据集包含 461 张手动标注有三个类别的腹腔镜图像:子宫、卵巢和手术工具。我们将数据集分为 361 张图像用于训练 Mask R-CNN,以及 100 张图像用于评估其性能。
分割准确性以 Mask R-CNN 分割区域与手动标注区域之间的重叠百分比来报告。子宫、卵巢和手术工具的准确性分别为 84.5%、29.6%和 54.5%。然后,从语义分割结果推断出这些结构的自动检测,这导致了除卵巢外的最新检测性能。具体来说,子宫、卵巢和手术工具的检测准确性分别为 97%、24%和 86%。
鉴于我们初始数据集的相对较小规模,我们的初步结果非常有希望。创建一个国际手术数据库似乎至关重要。