Suppr超能文献

评估单阶段视觉模型在手术器械位姿估计中的应用。

Evaluation of single-stage vision models for pose estimation of surgical instruments.

机构信息

Center for Orthopaedic Biomechanics, University of Denver, 2155 E Wesley Ave, Denver, CO, 80210, USA.

Unmanned Systems Research Institute, University of Denver, 2155 E Wesley Ave, Denver, CO, 80210, USA.

出版信息

Int J Comput Assist Radiol Surg. 2023 Dec;18(12):2125-2142. doi: 10.1007/s11548-023-02890-6. Epub 2023 Apr 30.

Abstract

PURPOSE

Multiple applications in open surgical environments may benefit from adoption of markerless computer vision depending on associated speed and accuracy requirements. The current work evaluates vision models for 6-degree of freedom pose estimation of surgical instruments in RGB scenes. Potential use cases are discussed based on observed performance.

METHODS

Convolutional neural nets were developed with simulated training data for 6-degree of freedom pose estimation of a representative surgical instrument in RGB scenes. Trained models were evaluated with simulated and real-world scenes. Real-world scenes were produced by using a robotic manipulator to procedurally generate a wide range of object poses.

RESULTS

CNNs trained in simulation transferred to real-world evaluation scenes with a mild decrease in pose accuracy. Model performance was sensitive to input image resolution and orientation prediction format. The model with highest accuracy demonstrated mean in-plane translation error of 13 mm and mean long axis orientation error of 5[Formula: see text] in simulated evaluation scenes. Similar errors of 29 mm and 8[Formula: see text] were observed in real-world scenes.

CONCLUSION

6-DoF pose estimators can predict object pose in RGB scenes with real-time inference speed. Observed pose accuracy suggests that applications such as coarse-grained guidance, surgical skill evaluation, or instrument tracking for tray optimization may benefit from markerless pose estimation.

摘要

目的

基于相关的速度和精度要求,无标记计算机视觉在开放式手术环境中的多种应用可能会受益于此。目前的工作评估了用于在 RGB 场景中对手术器械进行六自由度位姿估计的视觉模型。根据观察到的性能讨论了潜在的用例。

方法

针对 RGB 场景中代表性手术器械的六自由度位姿估计,使用模拟训练数据开发了卷积神经网络。使用模拟和真实场景评估了经过训练的模型。通过使用机器人操纵器来程序化地生成广泛的物体姿态,生成了真实场景。

结果

在模拟中训练的 CNN 转移到真实世界的评估场景时,位姿精度略有下降。模型性能对输入图像分辨率和方位预测格式敏感。在模拟评估场景中,具有最高精度的模型表现出平均面内平移误差为 13mm,平均长轴方位误差为 5[公式:见文本]。在真实场景中观察到类似的误差为 29mm 和 8[公式:见文本]。

结论

六自由度位姿估计器可以以实时推断速度预测 RGB 场景中的物体位姿。观察到的位姿精度表明,诸如粗粒度引导、手术技能评估或托盘优化的器械跟踪等应用可能受益于无标记位姿估计。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验