Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, No. 88 Keling Road, Suzhou New District, Suzhou, 215163, Jiangsu, China.
Jinan Guoke Medical Engineering and Technology Development Co., Ltd., No.3 Building, Pharmaceutical Valley New Drug Creation Platform, Jinan New District, Jinan, 250101, Shandong, China.
Int J Comput Assist Radiol Surg. 2020 Aug;15(8):1335-1345. doi: 10.1007/s11548-020-02214-y. Epub 2020 Jun 24.
The surgical instrument tracking framework, especially the marker-free surgical instrument tracking framework, is the key to visual servoing which is applied to achieve active control for laparoscope-holder robots. This paper presented a marker-free surgical instrument tracking framework based on object extraction via deep learning (DL).
The surgical instrument joint was defined as the tracking point. Using DL, a segmentation model was trained to extract the end-effector and shaft portions of the surgical instrument in real time. The extracted object was transformed into a distance image by Euclidean Distance Transformation. Next, the points with the maximal pixel value in the two portions were defined as their central points, respectively, and the intersection point of the line connecting the two central points and the plane connecting the two portions was determined as the tracking point. Finally, the object could be fast extracted using the masking method, and the tracking point was fast located frame-by-frame in a laparoscopic video to achieve tracking of surgical instrument. The proposed object extraction via a DL-based marker-free tracking framework was compared with a marker-free tracking-by-detection framework via DL.
Using seven in vivo laparoscopic videos for experiments, the mean tracking success rate was 100%. The mean tracking accuracy was (3.9 ± 2.4, 4.0 ± 2.5) pixels measured in u and v coordinates of a frame, and the mean tracking speed was 15 fps. Compared to the reported mean tracking accuracy of a marker-free tracking-by-detection framework via DL, the mean tracking accuracy of our proposed tracking framework was improved by 37% and 23%, respectively.
Accurate and fast tracking of marker-free surgical instruments could be achieved in in vivo laparoscopic videos by using our proposed object extraction via DL-based marker-free tracking framework. This work provided important guiding significance for the application of laparoscope-holder robots in laparoscopic surgeries.
手术器械跟踪框架,特别是无标记手术器械跟踪框架,是视觉伺服的关键,视觉伺服用于实现腹腔镜持镜机器人的主动控制。本文提出了一种基于深度学习(DL)的物体提取的无标记手术器械跟踪框架。
将手术器械关节定义为跟踪点。使用 DL 训练分割模型,实时提取手术器械的末端执行器和轴部分。提取的物体通过欧几里得距离变换转换为距离图像。接下来,将两个部分中具有最大像素值的点分别定义为它们的中心点,并确定连接两个中心点的线和连接两个部分的平面的交点作为跟踪点。最后,使用掩模法快速提取物体,在腹腔镜视频中逐帧快速定位跟踪点,实现手术器械的跟踪。将基于 DL 的无标记跟踪框架的物体提取与基于 DL 的无标记跟踪检测框架进行了比较。
使用 7 段体内腹腔镜视频进行实验,平均跟踪成功率为 100%。在一个帧的 u 和 v 坐标中,平均跟踪精度为(3.9±2.4,4.0±2.5)像素,平均跟踪速度为 15 fps。与已报道的基于 DL 的无标记跟踪检测框架的平均跟踪精度相比,我们提出的跟踪框架的平均跟踪精度分别提高了 37%和 23%。
通过使用基于 DL 的无标记跟踪框架的物体提取,可以在体内腹腔镜视频中实现无标记手术器械的准确快速跟踪。这项工作为腹腔镜持镜机器人在腹腔镜手术中的应用提供了重要的指导意义。