• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

从手术视频中探索在线相位识别的分段级语义。

Exploring Segment-Level Semantics for Online Phase Recognition From Surgical Videos.

出版信息

IEEE Trans Med Imaging. 2022 Nov;41(11):3309-3319. doi: 10.1109/TMI.2022.3182995. Epub 2022 Oct 27.

DOI:10.1109/TMI.2022.3182995
PMID:35700259
Abstract

Automatic surgical phase recognition plays a vital role in robot-assisted surgeries. Existing methods ignored a pivotal problem that surgical phases should be classified by learning segment-level semantics instead of solely relying on frame-wise information. This paper presents a segment-attentive hierarchical consistency network (SAHC) for surgical phase recognition from videos. The key idea is to extract hierarchical high-level semantic-consistent segments and use them to refine the erroneous predictions caused by ambiguous frames. To achieve it, we design a temporal hierarchical network to generate hierarchical high-level segments. Then, we introduce a hierarchical segment-frame attention module to capture relations between the low-level frames and high-level segments. By regularizing the predictions of frames and their corresponding segments via a consistency loss, the network can generate semantic-consistent segments and then rectify the misclassified predictions caused by ambiguous low-level frames. We validate SAHC on two public surgical video datasets, i.e., the M2CAI16 challenge dataset and the Cholec80 dataset. Experimental results show that our method outperforms previous state-of-the-arts and ablation studies prove the effectiveness of our proposed modules. Our code has been released at: https://github.com/xmed-lab/SAHC.

摘要

自动手术阶段识别在机器人辅助手术中起着至关重要的作用。现有的方法忽略了一个关键问题,即手术阶段应该通过学习段级别的语义进行分类,而不仅仅依赖于逐帧的信息。本文提出了一种用于从视频中识别手术阶段的基于分段注意力的层次一致性网络(SAHC)。其核心思想是提取分层的高级语义一致段,并利用它们来细化由于模糊帧引起的错误预测。为了实现这一点,我们设计了一个时间层次网络来生成分层的高级段。然后,我们引入了一个层次分段-帧注意力模块来捕获低级帧和高级段之间的关系。通过一致性损失来正则化帧及其相应段的预测,网络可以生成语义一致的段,然后纠正由于低级模糊帧引起的错误分类预测。我们在两个公共的手术视频数据集上验证了 SAHC,即 M2CAI16 挑战赛数据集和 Cholec80 数据集。实验结果表明,我们的方法优于以前的最新技术,并且消融研究证明了我们提出的模块的有效性。我们的代码已经发布在:https://github.com/xmed-lab/SAHC。

相似文献

1
Exploring Segment-Level Semantics for Online Phase Recognition From Surgical Videos.从手术视频中探索在线相位识别的分段级语义。
IEEE Trans Med Imaging. 2022 Nov;41(11):3309-3319. doi: 10.1109/TMI.2022.3182995. Epub 2022 Oct 27.
2
Semantic-Preserving Surgical Video Retrieval With Phase and Behavior Coordinated Hashing.基于相位和行为协调哈希的语义保持手术视频检索。
IEEE Trans Med Imaging. 2024 Feb;43(2):807-819. doi: 10.1109/TMI.2023.3321382. Epub 2024 Feb 2.
3
LoViT: Long Video Transformer for surgical phase recognition.LoViT:用于手术阶段识别的长视频 Transformer。
Med Image Anal. 2025 Jan;99:103366. doi: 10.1016/j.media.2024.103366. Epub 2024 Oct 5.
4
Global-local multi-stage temporal convolutional network for cataract surgery phase recognition.用于白内障手术阶段识别的全局-局部多阶段时间卷积网络。
Biomed Eng Online. 2022 Nov 30;21(1):82. doi: 10.1186/s12938-022-01048-w.
5
Exploring Intra- and Inter-Video Relation for Surgical Semantic Scene Segmentation.探索手术语义场景分割中的视频内和视频间关系。
IEEE Trans Med Imaging. 2022 Nov;41(11):2991-3002. doi: 10.1109/TMI.2022.3177077. Epub 2022 Oct 27.
6
SF-TMN: SlowFast temporal modeling network for surgical phase recognition.SF-TMN:用于手术阶段识别的慢快时变建模网络。
Int J Comput Assist Radiol Surg. 2024 May;19(5):871-880. doi: 10.1007/s11548-024-03095-1. Epub 2024 Mar 21.
7
LRTD: long-range temporal dependency based active learning for surgical workflow recognition.基于长程时间依赖的主动学习在手术流程识别中的应用
Int J Comput Assist Radiol Surg. 2020 Sep;15(9):1573-1584. doi: 10.1007/s11548-020-02198-9. Epub 2020 Jun 25.
8
LAST: LAtent Space-Constrained Transformers for Automatic Surgical Phase Recognition and Tool Presence Detection.LAST:用于自动手术阶段识别和工具存在检测的潜在空间约束式变换模型。
IEEE Trans Med Imaging. 2023 Nov;42(11):3256-3268. doi: 10.1109/TMI.2023.3279838. Epub 2023 Oct 27.
9
A spatio-temporal network for video semantic segmentation in surgical videos.用于手术视频中视频语义分割的时空网络。
Int J Comput Assist Radiol Surg. 2024 Feb;19(2):375-382. doi: 10.1007/s11548-023-02971-6. Epub 2023 Jun 22.
10
End-to-End Pre-Training With Hierarchical Matching and Momentum Contrast for Text-Video Retrieval.用于文本-视频检索的基于层次匹配和动量对比的端到端预训练
IEEE Trans Image Process. 2023;32:5017-5030. doi: 10.1109/TIP.2023.3275071. Epub 2023 Sep 8.

引用本文的文献

1
SWAG: long-term surgical workflow prediction with generative-based anticipation.SWAG:基于生成式预测的长期手术工作流程预测
Int J Comput Assist Radiol Surg. 2025 Jun 26. doi: 10.1007/s11548-025-03452-8.
2
Artificial intelligence-assisted phase recognition and skill assessment in laparoscopic surgery: a systematic review.腹腔镜手术中人工智能辅助的阶段识别与技能评估:一项系统综述
Front Surg. 2025 Apr 11;12:1551838. doi: 10.3389/fsurg.2025.1551838. eCollection 2025.
3
Augmenting efficient real-time surgical instrument segmentation in video with point tracking and Segment Anything.
利用点跟踪和“分割一切”模型增强视频中高效的实时手术器械分割
Healthc Technol Lett. 2024 Dec 30;12(1):e12111. doi: 10.1049/htl2.12111. eCollection 2025 Jan-Dec.
4
LoViT: Long Video Transformer for surgical phase recognition.LoViT:用于手术阶段识别的长视频 Transformer。
Med Image Anal. 2025 Jan;99:103366. doi: 10.1016/j.media.2024.103366. Epub 2024 Oct 5.
5
SF-TMN: SlowFast temporal modeling network for surgical phase recognition.SF-TMN:用于手术阶段识别的慢快时变建模网络。
Int J Comput Assist Radiol Surg. 2024 May;19(5):871-880. doi: 10.1007/s11548-024-03095-1. Epub 2024 Mar 21.
6
A Deep Learning Approach to Classify Surgical Skill in Microsurgery Using Force Data from a Novel Sensorised Surgical Glove.一种使用新型传感器手术手套力数据对微创手术中手术技能进行分类的深度学习方法。
Sensors (Basel). 2023 Nov 3;23(21):8947. doi: 10.3390/s23218947.
7
Automated segmentation of phases, steps, and tasks in laparoscopic cholecystectomy using deep learning.使用深度学习对腹腔镜胆囊切除术的阶段、步骤和任务进行自动分割。
Surg Endosc. 2024 Jan;38(1):158-170. doi: 10.1007/s00464-023-10482-3. Epub 2023 Nov 9.