Surgical Device Innovation Office, National Cancer Center Hospital East, 6-5-1 Kashiwanoha, Kashiwa, Chiba, 277-8577, Japan; Department of Colorectal Surgery, National Cancer Center Hospital East, 6-5-1 Kashiwanoha, Kashiwa, Chiba, 277-8577, Japan; Department of Gastrointestinal and Hepato-Biliary-Pancreatic Surgery, Faculty of Medicine, University of Tsukuba, 1-1-1 Tennodai, Tsukuba, Ibaraki, 305-8575, Japan.
Surgical Device Innovation Office, National Cancer Center Hospital East, 6-5-1 Kashiwanoha, Kashiwa, Chiba, 277-8577, Japan; Department of Colorectal Surgery, National Cancer Center Hospital East, 6-5-1 Kashiwanoha, Kashiwa, Chiba, 277-8577, Japan.
Int J Surg. 2020 Jul;79:88-94. doi: 10.1016/j.ijsu.2020.05.015. Epub 2020 May 12.
Identifying laparoscopic surgical videos using artificial intelligence (AI) facilitates the automation of several currently time-consuming manual processes, including video analysis, indexing, and video-based skill assessment. This study aimed to construct a large annotated dataset comprising laparoscopic colorectal surgery (LCRS) videos from multiple institutions and evaluate the accuracy of automatic recognition for surgical phase, action, and tool by combining this dataset with AI.
A total of 300 intraoperative videos were collected from 19 high-volume centers. A series of surgical workflows were classified into 9 phases and 3 actions, and the area of 5 tools were assigned by painting. More than 82 million frames were annotated for a phase and action classification task, and 4000 frames were annotated for a tool segmentation task. Of these frames, 80% were used for the training dataset and 20% for the test dataset. A convolutional neural network (CNN) was used to analyze the videos. Intersection over union (IoU) was used as the evaluation metric for tool recognition.
The overall accuracies for the automatic surgical phase and action classification task were 81.0% and 83.2%, respectively. The mean IoU for the automatic tool segmentation task for 5 tools was 51.2%.
A large annotated dataset of LCRS videos was constructed, and the phase, action, and tool were recognized with high accuracy using AI. Our dataset has potential uses in medical applications such as automatic video indexing and surgical skill assessments. Open research will assist in improving CNN models by making our dataset available in the field of computer vision.
使用人工智能 (AI) 识别腹腔镜手术视频有助于实现包括视频分析、索引和基于视频的技能评估在内的几个当前耗时的手动过程的自动化。本研究旨在构建一个包含来自多个机构的腹腔镜结直肠手术 (LCRS) 视频的大型标注数据集,并通过将该数据集与 AI 相结合,评估自动识别手术阶段、动作和工具的准确性。
从 19 个大容量中心共采集了 300 个术中视频。一系列手术工作流程分为 9 个阶段和 3 个动作,并通过绘画分配 5 个工具的区域。为阶段和动作分类任务标注了超过 8200 万帧,为工具分割任务标注了 4000 帧。其中 80%的帧用于训练数据集,20%的帧用于测试数据集。使用卷积神经网络 (CNN) 分析视频。交并比 (IoU) 用作工具识别的评估指标。
自动手术阶段和动作分类任务的总体准确率分别为 81.0%和 83.2%。5 个工具的自动工具分割任务的平均 IoU 为 51.2%。
构建了一个大型的 LCRS 视频标注数据集,并使用 AI 实现了高准确度的阶段、动作和工具识别。我们的数据集在医疗应用中具有潜力,例如自动视频索引和手术技能评估。开放研究将通过在计算机视觉领域提供我们的数据集来协助改进 CNN 模型。