Harvard University, Cambridge, MA.
Department of Automation, Tsinghua University, Beijing, China.
AMIA Annu Symp Proc. 2021 Jan 25;2020:1373-1382. eCollection 2020.
Open, or non-laparoscopic surgery, represents the vast majority of all operating room procedures, but few tools exist to objectively evaluate these techniques at scale. Current efforts involve human expert-based visual assessment. We leverage advances in computer vision to introduce an automated approach to video analysis of surgical execution. A state-of-the-art convolutional neural network architecture for object detection was used to detect operating hands in open surgery videos. Automated assessment was expanded by combining model predictions with a fast object tracker to enable surgeon-specific hand tracking. To train our model, we used publicly available videos of open surgery from YouTube and annotated these with spatial bounding boxes of operating hands. Our model's spatial detections of operating hands significantly outperforms the detections achieved using pre-existing hand-detection datasets, and allow for insights into intra-operative movement patterns and economy of motion.
开放式手术(非腹腔镜手术)代表了绝大多数手术室手术,但目前几乎没有客观评估这些技术的工具。目前的研究工作涉及基于人类专家的视觉评估。我们利用计算机视觉的最新进展,引入了一种自动的手术执行视频分析方法。我们使用了一种最先进的用于目标检测的卷积神经网络架构,来检测开放式手术视频中的手术手。通过将模型预测与快速目标跟踪器相结合,实现了针对特定外科医生的手部跟踪,从而扩展了自动化评估。为了训练我们的模型,我们使用了来自 YouTube 的公开可用的开放式手术视频,并使用空间边界框对手术手进行了注释。我们的模型对手部的空间检测明显优于使用现有手部检测数据集所达到的检测效果,并能够深入了解手术过程中的运动模式和运动经济性。