Department of Computing and Mathematical Sciences, California Institute of Technology, Pasadena, CA, USA.
Center for Robotic Simulation and Education, Catherine & Joseph Aresty Department of Urology, University of Southern California, Los Angeles, CA, USA.
Nat Biomed Eng. 2023 Jun;7(6):780-796. doi: 10.1038/s41551-023-01010-8. Epub 2023 Mar 30.
The intraoperative activity of a surgeon has substantial impact on postoperative outcomes. However, for most surgical procedures, the details of intraoperative surgical actions, which can vary widely, are not well understood. Here we report a machine learning system leveraging a vision transformer and supervised contrastive learning for the decoding of elements of intraoperative surgical activity from videos commonly collected during robotic surgeries. The system accurately identified surgical steps, actions performed by the surgeon, the quality of these actions and the relative contribution of individual video frames to the decoding of the actions. Through extensive testing on data from three different hospitals located in two different continents, we show that the system generalizes across videos, surgeons, hospitals and surgical procedures, and that it can provide information on surgical gestures and skills from unannotated videos. Decoding intraoperative activity via accurate machine learning systems could be used to provide surgeons with feedback on their operating skills, and may allow for the identification of optimal surgical behaviour and for the study of relationships between intraoperative factors and postoperative outcomes.
外科医生的术中活动对术后结果有重大影响。然而,对于大多数外科手术,其术中手术操作的细节(差异很大)尚不清楚。在这里,我们报告了一个利用视觉转换器和监督对比学习的机器学习系统,用于从机器人手术中通常采集的视频中解码术中外科活动的元素。该系统能够准确识别手术步骤、外科医生执行的操作、这些操作的质量以及各个视频帧对操作解码的相对贡献。通过在来自两个不同大洲的三个不同医院的数据上进行广泛测试,我们表明该系统可以跨视频、外科医生、医院和手术程序进行泛化,并且可以从未标注的视频中提供有关手术手势和技能的信息。通过准确的机器学习系统对术中活动进行解码,可以为外科医生提供操作技能的反馈,并且可以识别最佳手术行为,并研究术中因素与术后结果之间的关系。