Surgical Device Innovation Office, National Cancer Center Hospital East, Kashiwanoha, Kashiwa, Chiba, Japan.
Department of Colorectal Surgery, National Cancer Center Hospital East, Kashiwanoha, Kashiwa, Chiba, Japan.
JAMA Netw Open. 2022 Aug 1;5(8):e2226265. doi: 10.1001/jamanetworkopen.2022.26265.
Deep learning-based automatic surgical instrument recognition is an indispensable technology for surgical research and development. However, pixel-level recognition with high accuracy is required to make it suitable for surgical automation.
To develop a deep learning model that can simultaneously recognize 8 types of surgical instruments frequently used in laparoscopic colorectal operations and evaluate its recognition performance.
DESIGN, SETTING, AND PARTICIPANTS: This quality improvement study was conducted at a single institution with a multi-institutional data set. Laparoscopic colorectal surgical videos recorded between April 1, 2009, and December 31, 2021, were included in the video data set. Deep learning-based instance segmentation, an image recognition approach that recognizes each object individually and pixel by pixel instead of roughly enclosing with a bounding box, was performed for 8 types of surgical instruments.
Average precision, calculated from the area under the precision-recall curve, was used as an evaluation metric. The average precision represents the number of instances of true-positive, false-positive, and false-negative results, and the mean average precision value for 8 types of surgical instruments was calculated. Five-fold cross-validation was used as the validation method. The annotation data set was split into 5 segments, of which 4 were used for training and the remainder for validation. The data set was split at the per-case level instead of the per-frame level; thus, the images extracted from an intraoperative video in the training set never appeared in the validation set. Validation was performed for all 5 validation sets, and the average mean average precision was calculated.
In total, 337 laparoscopic colorectal surgical videos were used. Pixel-by-pixel annotation was manually performed for 81 760 labels on 38 628 static images, constituting the annotation data set. The mean average precisions of the instance segmentation for surgical instruments were 90.9% for 3 instruments, 90.3% for 4 instruments, 91.6% for 6 instruments, and 91.8% for 8 instruments.
A deep learning-based instance segmentation model that simultaneously recognizes 8 types of surgical instruments with high accuracy was successfully developed. The accuracy was maintained even when the number of types of surgical instruments increased. This model can be applied to surgical innovations, such as intraoperative navigation and surgical automation.
基于深度学习的自动手术器械识别是手术研究和开发不可或缺的技术。然而,要使其适用于手术自动化,需要进行高精度的像素级识别。
开发一种能够同时识别腹腔镜结直肠手术中常用的 8 种手术器械的深度学习模型,并评估其识别性能。
设计、设置和参与者:这是一项单机构多机构数据集的质量改进研究。纳入了 2009 年 4 月 1 日至 2021 年 12 月 31 日期间记录的腹腔镜结直肠手术视频的视频数据集。对 8 种手术器械进行了基于深度学习的实例分割,这是一种图像识别方法,它逐个像素地识别每个对象,而不是用边界框粗略地包围。
使用精度-召回曲线下面积计算的平均精度作为评估指标。平均精度表示真阳性、假阳性和假阴性结果的数量,并计算了 8 种手术器械的平均精度值。使用 5 折交叉验证作为验证方法。注释数据集被分为 5 个部分,其中 4 个用于训练,其余部分用于验证。数据集是在逐例的基础上而不是逐帧的基础上进行分割的;因此,从训练集中的一个术中视频提取的图像从未出现在验证集中。对所有 5 个验证集进行验证,并计算平均平均精度。
共使用了 337 个腹腔镜结直肠手术视频。在 38628 张静态图像上对 81760 个标签进行了逐像素注释,构成了注释数据集。手术器械实例分割的平均精度分别为 3 种器械 90.9%、4 种器械 90.3%、6 种器械 91.6%和 8 种器械 91.8%。
成功开发了一种能够高精度同时识别 8 种手术器械的基于深度学习的实例分割模型。即使增加了手术器械的种类,其准确性也得以保持。该模型可应用于手术创新,如术中导航和手术自动化。