• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

WHSP-Net:一种用于从单张深度图像中恢复三维手部形状和姿态的弱监督方法。

WHSP-Net: A Weakly-Supervised Approach for 3D Hand Shape and Pose Recovery from a Single Depth Image.

作者信息

Malik Jameel, Elhayek Ahmed, Stricker Didier

机构信息

German Research Center for Artificial Intelligence, DFKI, 67663 Kaiserslautern, Germany.

Department of Informatics, University of Kaiserslautern, 67653 Kaiserslautern, Germany.

出版信息

Sensors (Basel). 2019 Aug 31;19(17):3784. doi: 10.3390/s19173784.

DOI:10.3390/s19173784
PMID:31480461
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC6749339/
Abstract

Hand shape and pose recovery is essential for many computer vision applications such as animation of a personalized hand mesh in a virtual environment. Although there are many hand pose estimation methods, only a few deep learning based algorithms target 3D hand shape and pose from a single RGB or depth image. Jointly estimating hand shape and pose is very challenging because none of the existing real benchmarks provides ground truth hand shape. For this reason, we propose a novel weakly-supervised approach for 3D hand shape and pose recovery (named WHSP-Net) from a single depth image by learning shapes from unlabeled real data and labeled synthetic data. To this end, we propose a novel framework which consists of three novel components. The first is the Convolutional Neural Network (CNN) based deep network which produces 3D joints positions from learned 3D bone vectors using a new layer. The second is a novel shape decoder that recovers dense 3D hand mesh from sparse joints. The third is a novel depth synthesizer which reconstructs 2D depth image from 3D hand mesh. The whole pipeline is fine-tuned in an end-to-end manner. We demonstrate that our approach recovers reasonable hand shapes from real world datasets as well as from live stream of depth camera in real-time. Our algorithm outperforms state-of-the-art methods that output more than the joint positions and shows competitive performance on 3D pose estimation task.

摘要

手部形状和姿态恢复对于许多计算机视觉应用至关重要,例如在虚拟环境中对个性化手部网格进行动画处理。尽管有许多手部姿态估计方法,但只有少数基于深度学习的算法能够从单张RGB或深度图像中确定3D手部形状和姿态。联合估计手部形状和姿态极具挑战性,因为现有的真实基准测试均未提供真实的手部形状。因此,我们提出了一种新颖的弱监督方法,用于从单张深度图像中恢复3D手部形状和姿态(名为WHSP-Net),该方法通过从未标记的真实数据和标记的合成数据中学习形状来实现。为此,我们提出了一个新颖的框架,它由三个新颖的组件组成。第一个是基于卷积神经网络(CNN)的深度网络,该网络使用一个新层从学习到的3D骨骼向量中生成3D关节位置。第二个是新颖的形状解码器,它从稀疏关节中恢复密集的3D手部网格。第三个是新颖的深度合成器,它从3D手部网格重建2D深度图像。整个管道以端到端的方式进行微调。我们证明,我们的方法能够从真实世界数据集以及深度相机的实时流中实时恢复合理的手部形状。我们的算法优于那些输出超过关节位置的现有方法,并在3D姿态估计任务中表现出具有竞争力的性能。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/4f78c3eeefd5/sensors-19-03784-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/34f9ac134259/sensors-19-03784-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/f5870f6b165f/sensors-19-03784-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/c4027c411b22/sensors-19-03784-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/fe0a5f3948d3/sensors-19-03784-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/4e00dcb315f7/sensors-19-03784-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/2483be4503ed/sensors-19-03784-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/c9ffd80071e2/sensors-19-03784-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/5f237c7a34da/sensors-19-03784-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/814c08b40a20/sensors-19-03784-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/9a0c180a2776/sensors-19-03784-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/4f78c3eeefd5/sensors-19-03784-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/34f9ac134259/sensors-19-03784-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/f5870f6b165f/sensors-19-03784-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/c4027c411b22/sensors-19-03784-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/fe0a5f3948d3/sensors-19-03784-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/4e00dcb315f7/sensors-19-03784-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/2483be4503ed/sensors-19-03784-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/c9ffd80071e2/sensors-19-03784-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/5f237c7a34da/sensors-19-03784-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/814c08b40a20/sensors-19-03784-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/9a0c180a2776/sensors-19-03784-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9586/6749339/4f78c3eeefd5/sensors-19-03784-g011.jpg

相似文献

1
WHSP-Net: A Weakly-Supervised Approach for 3D Hand Shape and Pose Recovery from a Single Depth Image.WHSP-Net:一种用于从单张深度图像中恢复三维手部形状和姿态的弱监督方法。
Sensors (Basel). 2019 Aug 31;19(17):3784. doi: 10.3390/s19173784.
2
Weakly Supervised Learning for Single Depth-Based Hand Shape Recovery.基于单目深度图的弱监督手形恢复。
IEEE Trans Image Process. 2021;30:532-545. doi: 10.1109/TIP.2020.3037479. Epub 2020 Nov 24.
3
Real-Time 3D Hand Pose Estimation with 3D Convolutional Neural Networks.基于3D卷积神经网络的实时3D手部姿态估计
IEEE Trans Pattern Anal Mach Intell. 2019 Apr;41(4):956-970. doi: 10.1109/TPAMI.2018.2827052. Epub 2018 Apr 16.
4
3D Hand Pose Estimation Using Synthetic Data and Weakly Labeled RGB Images.基于合成数据和弱标注 RGB 图像的三维手姿估计。
IEEE Trans Pattern Anal Mach Intell. 2021 Nov;43(11):3739-3753. doi: 10.1109/TPAMI.2020.2993627. Epub 2021 Oct 1.
5
Weakly Supervised Adversarial Learning for 3D Human Pose Estimation from Point Clouds.基于点云的弱监督对抗学习三维人体姿态估计
IEEE Trans Vis Comput Graph. 2020 May;26(5):1851-1859. doi: 10.1109/TVCG.2020.2973076. Epub 2020 Feb 13.
6
HandVoxNet++: 3D Hand Shape and Pose Estimation Using Voxel-Based Neural Networks.HandVoxNet++:基于体素的神经网络的三维手形状和姿态估计。
IEEE Trans Pattern Anal Mach Intell. 2022 Dec;44(12):8962-8974. doi: 10.1109/TPAMI.2021.3122874. Epub 2022 Nov 7.
7
3D hand pose and mesh estimation via a generic Topology-aware Transformer model.通过通用拓扑感知变压器模型进行3D手部姿态和网格估计。
Front Neurorobot. 2024 May 3;18:1395652. doi: 10.3389/fnbot.2024.1395652. eCollection 2024.
8
Fusing information from multiple 2D depth cameras for 3D human pose estimation in the operating room.将来自多个 2D 深度相机的信息融合用于手术室中的 3D 人体姿态估计。
Int J Comput Assist Radiol Surg. 2019 Nov;14(11):1871-1879. doi: 10.1007/s11548-019-02044-7. Epub 2019 Aug 6.
9
Robust 3D Hand Pose Estimation From Single Depth Images Using Multi-View CNNs.基于多视角卷积神经网络的单目深度图像稳健三维手姿估计
IEEE Trans Image Process. 2018 Sep;27(9):4422-4436. doi: 10.1109/TIP.2018.2834824.
10
Generalized Feedback Loop for Joint Hand-Object Pose Estimation.通用的手-物联合位姿估计反馈环。
IEEE Trans Pattern Anal Mach Intell. 2020 Aug;42(8):1898-1912. doi: 10.1109/TPAMI.2019.2907951. Epub 2019 Mar 27.

引用本文的文献

1
Intelligent Tensioning Method for Prestressed Cables Based on Digital Twins and Artificial Intelligence.基于数字孪生和人工智能的预应力索智能张拉方法
Sensors (Basel). 2020 Dec 8;20(24):7006. doi: 10.3390/s20247006.
2
Structure from Articulated Motion: Accurate and Stable Monocular 3D Reconstruction without Training Data.从关节运动中构建结构:无需训练数据的准确且稳定的单目 3D 重建。
Sensors (Basel). 2019 Oct 22;19(20):4603. doi: 10.3390/s19204603.

本文引用的文献

1
Context-Aware Deep Spatiotemporal Network for Hand Pose Estimation From Depth Images.基于深度图像的上下文感知深度时空网络的手势估计。
IEEE Trans Cybern. 2020 Feb;50(2):787-797. doi: 10.1109/TCYB.2018.2873733. Epub 2018 Dec 5.
2
3DAirSig: A Framework for Enabling In-Air Signatures Using a Multi-Modal Depth Sensor.3DAirSig:一种使用多模态深度传感器实现空中签名的框架。
Sensors (Basel). 2018 Nov 10;18(11):3872. doi: 10.3390/s18113872.
3
Deep learning in neural networks: an overview.神经网络中的深度学习:综述。
Neural Netw. 2015 Jan;61:85-117. doi: 10.1016/j.neunet.2014.09.003. Epub 2014 Oct 13.