• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

将来自多个 2D 深度相机的信息融合用于手术室中的 3D 人体姿态估计。

Fusing information from multiple 2D depth cameras for 3D human pose estimation in the operating room.

机构信息

Institute of Medical Informatics, University of Lübeck, Ratzeburger Allee 160, 23562, Lübeck, Germany.

Drägerwerk AG & Co. KGaA, Moislinger Allee 53-55, 23558, Lübeck, Germany.

出版信息

Int J Comput Assist Radiol Surg. 2019 Nov;14(11):1871-1879. doi: 10.1007/s11548-019-02044-7. Epub 2019 Aug 6.

DOI:10.1007/s11548-019-02044-7
PMID:31388959
Abstract

PURPOSE

For many years, deep convolutional neural networks have achieved state-of-the-art results on a wide variety of computer vision tasks. 3D human pose estimation makes no exception and results on public benchmarks are impressive. However, specialized domains, such as operating rooms, pose additional challenges. Clinical settings include severe occlusions, clutter and difficult lighting conditions. Privacy concerns of patients and staff make it necessary to use unidentifiable data. In this work, we aim to bring robust human pose estimation to the clinical domain.

METHODS

We propose a 2D-3D information fusion framework that makes use of a network of multiple depth cameras and strong pose priors. In a first step, probabilities of 2D joints are predicted from single depth images. These information are fused in a shared voxel space yielding a rough estimate of the 3D pose. Final joint positions are obtained by regressing into the latent pose space of a pre-trained convolutional autoencoder.

RESULTS

We evaluate our approach against several baselines on the challenging MVOR dataset. Best results are obtained when fusing 2D information from multiple views and constraining the predictions with learned pose priors.

CONCLUSIONS

We present a robust 3D human pose estimation framework based on a multi-depth camera network in the operating room. Depth images as only input modalities make our approach especially interesting for clinical applications due to the given anonymity for patients and staff.

摘要

目的

多年来,深度卷积神经网络在各种计算机视觉任务上取得了最先进的成果。3D 人体姿态估计也不例外,在公共基准测试上的结果令人印象深刻。然而,特殊领域,如手术室,带来了额外的挑战。临床环境包括严重的遮挡、杂乱和困难的照明条件。患者和工作人员的隐私问题使得使用无法识别的数据成为必要。在这项工作中,我们旨在将强大的人体姿态估计引入临床领域。

方法

我们提出了一种 2D-3D 信息融合框架,该框架利用了多个深度摄像机网络和强大的姿态先验。在第一步中,从单个深度图像预测 2D 关节的概率。这些信息在共享体素空间中融合,得出 3D 姿态的粗略估计。最终的关节位置通过回归到预先训练的卷积自动编码器的潜在姿态空间来获得。

结果

我们在具有挑战性的 MVOR 数据集上对几种基线方法进行了评估。当融合来自多个视图的 2D 信息并利用学习到的姿态先验来约束预测时,会得到最佳结果。

结论

我们提出了一种基于手术室中多深度摄像机网络的强大 3D 人体姿态估计框架。由于患者和工作人员的匿名性,深度图像作为唯一的输入模态,使我们的方法特别适用于临床应用。

相似文献

1
Fusing information from multiple 2D depth cameras for 3D human pose estimation in the operating room.将来自多个 2D 深度相机的信息融合用于手术室中的 3D 人体姿态估计。
Int J Comput Assist Radiol Surg. 2019 Nov;14(11):1871-1879. doi: 10.1007/s11548-019-02044-7. Epub 2019 Aug 6.
2
Robust 3D Hand Pose Estimation From Single Depth Images Using Multi-View CNNs.基于多视角卷积神经网络的单目深度图像稳健三维手姿估计
IEEE Trans Image Process. 2018 Sep;27(9):4422-4436. doi: 10.1109/TIP.2018.2834824.
3
WHSP-Net: A Weakly-Supervised Approach for 3D Hand Shape and Pose Recovery from a Single Depth Image.WHSP-Net:一种用于从单张深度图像中恢复三维手部形状和姿态的弱监督方法。
Sensors (Basel). 2019 Aug 31;19(17):3784. doi: 10.3390/s19173784.
4
Real-Time 3D Hand Pose Estimation with 3D Convolutional Neural Networks.基于3D卷积神经网络的实时3D手部姿态估计
IEEE Trans Pattern Anal Mach Intell. 2019 Apr;41(4):956-970. doi: 10.1109/TPAMI.2018.2827052. Epub 2018 Apr 16.
5
Center point to pose: Multiple views 3D human pose estimation for multi-person.中心点姿态:多人多角度三维人体姿态估计
PLoS One. 2022 Sep 13;17(9):e0274450. doi: 10.1371/journal.pone.0274450. eCollection 2022.
6
Weakly Supervised Adversarial Learning for 3D Human Pose Estimation from Point Clouds.基于点云的弱监督对抗学习三维人体姿态估计
IEEE Trans Vis Comput Graph. 2020 May;26(5):1851-1859. doi: 10.1109/TVCG.2020.2973076. Epub 2020 Feb 13.
7
Adversarial Learning of Structure-Aware Fully Convolutional Networks for Landmark Localization.用于地标定位的结构感知全卷积网络的对抗学习
IEEE Trans Pattern Anal Mach Intell. 2020 Jul;42(7):1654-1669. doi: 10.1109/TPAMI.2019.2901875. Epub 2019 Feb 26.
8
A deep learning approach for pose estimation from volumetric OCT data.基于深度学习的体 OCT 数据的姿态估计方法。
Med Image Anal. 2018 May;46:162-179. doi: 10.1016/j.media.2018.03.002. Epub 2018 Mar 10.
9
An Efficient 3D Human Pose Retrieval and Reconstruction from 2D Image-Based Landmarks.基于二维图像特征点的高效三维人体姿态检索与重建。
Sensors (Basel). 2021 Apr 1;21(7):2415. doi: 10.3390/s21072415.
10
Articulated clinician detection using 3D pictorial structures on RGB-D data.基于 RGB-D 数据的三维图像结构的关节式临床医生检测。
Med Image Anal. 2017 Jan;35:215-224. doi: 10.1016/j.media.2016.07.001. Epub 2016 Jul 11.

引用本文的文献

1
When do they StOP?: A first step toward automatically identifying team communication in the operating room.他们何时停止?:迈向自动识别手术室团队沟通的第一步。
Int J Comput Assist Radiol Surg. 2025 May 14. doi: 10.1007/s11548-025-03393-2.
2
Using Gesture and Speech to Control Surgical Lighting Systems: Mixed Methods Study.使用手势和语音控制手术照明系统:混合方法研究
JMIR Hum Factors. 2025 May 12;12:e70628. doi: 10.2196/70628.
3
NeRF-OR: neural radiance fields for operating room scene reconstruction from sparse-view RGB-D videos.

本文引用的文献

1
In-Bed Pose Estimation: Deep Learning With Shallow Dataset.床上姿势估计:基于浅层数据集的深度学习
IEEE J Transl Eng Health Med. 2019 Jan 14;7:4900112. doi: 10.1109/JTEHM.2019.2892970. eCollection 2019.
2
Patient-Specific Pose Estimation in Clinical Environments.临床环境中的个性化姿势估计
IEEE J Transl Eng Health Med. 2018 Oct 10;6:2101111. doi: 10.1109/JTEHM.2018.2875464. eCollection 2018.
3
Temporal Trends and Characteristics of Reportable Health Data Breaches, 2010-2017.2010-2017 年可报告健康数据泄露事件的时间趋势和特征。
NeRF-OR:用于从稀疏视图RGB-D视频重建手术室场景的神经辐射场
Int J Comput Assist Radiol Surg. 2025 Jan;20(1):147-156. doi: 10.1007/s11548-024-03261-5. Epub 2024 Sep 13.
4
Real-time deep learning semantic segmentation during intra-operative surgery for 3D augmented reality assistance.术中实时深度学习语义分割用于三维增强现实辅助手术。
Int J Comput Assist Radiol Surg. 2021 Sep;16(9):1435-1445. doi: 10.1007/s11548-021-02432-y. Epub 2021 Jun 24.
JAMA. 2018 Sep 25;320(12):1282-1284. doi: 10.1001/jama.2018.9222.
4
Human3.6M: Large Scale Datasets and Predictive Methods for 3D Human Sensing in Natural Environments.Human3.6M:自然环境中 3D 人体感应的大规模数据集和预测方法。
IEEE Trans Pattern Anal Mach Intell. 2014 Jul;36(7):1325-39. doi: 10.1109/TPAMI.2013.248.
5
Video recording of the operating room--is anonymity possible?手术室的视频记录——能否做到匿名?
J Surg Res. 2015 Aug;197(2):272-6. doi: 10.1016/j.jss.2015.03.097. Epub 2015 Apr 9.
6
Statistical modeling and recognition of surgical workflow.手术流程的统计建模与识别。
Med Image Anal. 2012 Apr;16(3):632-41. doi: 10.1016/j.media.2010.10.001. Epub 2010 Dec 8.