Jia Tingting, Taylor Zeike A, Chen Xiaojun
School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, China; Institute of Medical Robotics, Shanghai Jiao Tong University, 200240, Shanghai, China.
CISTIB Centre for Computational Imaging and Simulation Technologies in Biomedicine, Institute of Medical and Biological Engineering, University of Leeds, Leeds, UK.
Comput Med Imaging Graph. 2021 Dec;94:101995. doi: 10.1016/j.compmedimag.2021.101995. Epub 2021 Oct 2.
Real-time augmented reality (AR) for minimally invasive surgery without extra tracking devices is a valuable yet challenging task, especially considering dynamic surgery environments. Multiple different motions between target organs are induced by respiration, cardiac motion or operative tools, and often must be characterized by a moving, manually positioned endoscope. Therefore, a 6DoF motion tracking method that takes advantage of the latest 2D target tracking methods and non-linear pose optimization and tracking loss retrieval in SLAM technologies is proposed and can be embedded into such an AR system. Specifically, the SiamMask deep learning-based target tracking method is incorporated to roughly exclude motion distractions and enable frame matching. This algorithm's light computation cost makes it possible for the proposed method to run in real-time. A global map and a set of keyframes as in ORB-SLAM are maintained for pose optimization and tracking loss retrieval. The stereo matching and frame matching methods are improved and a new strategy to select reference frames is introduced to make the first-time motion estimation of every arriving frame as accurate as possible. Experiments on both a clinical laparoscopic partial nephrectomy dataset and an ex-vivo porcine kidney dataset are conducted. The results show that the proposed method gives a more robust and accurate performance compared with ORB-SLAM2 in the presence of motion distractions or motion blur; however, heavy smoke still remains a big factor that reduces the tracking accuracy.
无需额外跟踪设备的实时增强现实(AR)用于微创手术是一项有价值但具有挑战性的任务,尤其是考虑到动态手术环境。目标器官之间的多种不同运动是由呼吸、心脏运动或手术工具引起的,并且通常必须通过移动的、手动定位的内窥镜来表征。因此,提出了一种利用最新的二维目标跟踪方法以及SLAM技术中的非线性姿态优化和跟踪损失检索的六自由度运动跟踪方法,并可将其嵌入到这样的AR系统中。具体来说,采用基于深度学习的SiamMask目标跟踪方法来大致排除运动干扰并实现帧匹配。该算法的轻量级计算成本使得所提出的方法能够实时运行。像在ORB-SLAM中一样维护一个全局地图和一组关键帧,用于姿态优化和跟踪损失检索。改进了立体匹配和帧匹配方法,并引入了一种选择参考帧的新策略,以使每个到达帧的首次运动估计尽可能准确。在临床腹腔镜部分肾切除术数据集和离体猪肾数据集上都进行了实验。结果表明,在存在运动干扰或运动模糊的情况下,与ORB-SLAM2相比,所提出的方法具有更稳健和准确的性能;然而,浓烟仍然是降低跟踪精度的一个重要因素。