Deol Ekamjit S, Tollefson Matthew K, Antolin Alenka, Zohar Maya, Bar Omri, Ben-Ayoun Danielle, Mynderse Lance A, Lomas Derek J, Avant Ross A, Miller Adam R, Elliott Daniel S, Boorjian Stephen A, Wolf Tamir, Asselmann Dotan, Khanna Abhinav
Department of Urology, Mayo Clinic, Rochester, MN, United States.
theator.io, Palo Alto, CA, United States.
Front Artif Intell. 2024 Mar 7;7:1375482. doi: 10.3389/frai.2024.1375482. eCollection 2024.
Automated surgical step recognition (SSR) using AI has been a catalyst in the "digitization" of surgery. However, progress has been limited to laparoscopy, with relatively few SSR tools in endoscopic surgery. This study aimed to create a SSR model for transurethral resection of bladder tumors (TURBT), leveraging a novel application of transfer learning to reduce video dataset requirements.
Retrospective surgical videos of TURBT were manually annotated with the following steps of surgery: primary endoscopic evaluation, resection of bladder tumor, and surface coagulation. Manually annotated videos were then utilized to train a novel AI computer vision algorithm to perform automated video annotation of TURBT surgical video, utilizing a transfer-learning technique to pre-train on laparoscopic procedures. Accuracy of AI SSR was determined by comparison to human annotations as the reference standard.
A total of 300 full-length TURBT videos (median 23.96 min; IQR 14.13-41.31 min) were manually annotated with sequential steps of surgery. One hundred and seventy-nine videos served as a training dataset for algorithm development, 44 for internal validation, and 77 as a separate test cohort for evaluating algorithm accuracy. Overall accuracy of AI video analysis was 89.6%. Model accuracy was highest for the primary endoscopic evaluation step (98.2%) and lowest for the surface coagulation step (82.7%).
We developed a fully automated computer vision algorithm for high-accuracy annotation of TURBT surgical videos. This represents the first application of transfer-learning from laparoscopy-based computer vision models into surgical endoscopy, demonstrating the promise of this approach in adapting to new procedure types.
利用人工智能实现手术步骤自动识别(SSR)已成为手术“数字化”的催化剂。然而,进展仅限于腹腔镜手术,内镜手术中的SSR工具相对较少。本研究旨在创建一种用于经尿道膀胱肿瘤切除术(TURBT)的SSR模型,利用迁移学习的新应用来减少视频数据集需求。
对TURBT的回顾性手术视频进行手动标注,标注手术的以下步骤:初步内镜评估、膀胱肿瘤切除和表面凝固。然后利用手动标注的视频训练一种新型人工智能计算机视觉算法,以对TURBT手术视频进行自动视频标注,利用迁移学习技术在腹腔镜手术程序上进行预训练。以人工标注作为参考标准,通过比较来确定人工智能SSR的准确性。
共对300个全长TURBT视频(中位数23.96分钟;四分位间距14.13 - 41.31分钟)进行了手术步骤的顺序标注。179个视频用作算法开发的训练数据集,44个用于内部验证,77个作为单独的测试队列用于评估算法准确性。人工智能视频分析的总体准确率为89.6%。模型在初步内镜评估步骤的准确率最高(98.2%),在表面凝固步骤的准确率最低(82.7%)。
我们开发了一种用于TURBT手术视频高精度标注的全自动计算机视觉算法。这代表了基于腹腔镜的计算机视觉模型的迁移学习首次应用于手术内镜,证明了这种方法在适应新手术类型方面的前景。