Annu Int Conf IEEE Eng Med Biol Soc. 2021 Nov;2021:1824-1827. doi: 10.1109/EMBC46164.2021.9629914.
Image-based tracking of laparoscopic instruments plays a fundamental role in computer and robotic-assisted surgeries by aiding surgeons and increasing patient safety. Computer vision contests, such as the Robust Medical Instrument Segmentation (ROBUST-MIS) Challenge, seek to encourage the development of robust models for such purposes, providing large, diverse, and high-quality datasets. To date, most of the existing models for instance segmentation of medical instruments were based on two-stage detectors, which provide robust results but are nowhere near to the real-time, running at 5 frames-per-second (fps) at most. However, for the method to be clinically applicable, a real-time capability is utmost required along with high accuracy. In this paper, we propose the addition of attention mechanisms to the YOLACT architecture to allow real-time instance segmentation of instruments with improved accuracy on the ROBUST-MIS dataset. Our proposed approach achieves competitive performance compared to the winner of the 2019 ROBUST-MIS challenge in terms of robustness scores, obtaining 0.313 ML_DSC and 0.338 MLNSD while reaching real-time performance at >45 fps.
基于图像的腹腔镜器械跟踪在计算机辅助和机器人辅助手术中起着至关重要的作用,它可以帮助外科医生并提高患者安全性。计算机视觉竞赛,如 Robust Medical Instrument Segmentation (ROBUST-MIS) 挑战赛,旨在鼓励开发用于此类目的的强大模型,提供大型、多样化和高质量的数据集。迄今为止,大多数用于医学器械实例分割的现有模型都是基于两阶段探测器的,这些模型提供了强大的结果,但远非实时,最多只能达到 5 帧/秒 (fps)。然而,为了使该方法在临床上适用,还需要实时能力和高精度。在本文中,我们建议在 YOLACT 架构中添加注意力机制,以允许实时分割器械实例,并在 ROBUST-MIS 数据集上提高准确性。与 2019 年 ROBUST-MIS 挑战赛的获胜者相比,我们的方法在稳健性得分方面具有竞争力,获得了 0.313 ML_DSC 和 0.338 MLNSD,同时达到了 >45 fps 的实时性能。