• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

HandFI:基于RGB图像中多级别特征融合的多级别交互手部重建

HandFI: Multilevel Interacting Hand Reconstruction Based on Multilevel Feature Fusion in RGB Images.

作者信息

Pan Huimin, Cai Yuting, Yang Jiayi, Niu Shaojia, Gao Quanli, Wang Xihan

机构信息

School of Computer Science, Xi'an Polytechnic University, Xi'an 710600, China.

出版信息

Sensors (Basel). 2024 Dec 27;25(1):88. doi: 10.3390/s25010088.

DOI:10.3390/s25010088
PMID:39796887
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11722860/
Abstract

Interacting hand reconstruction presents significant opportunities in various applications. However, it currently faces challenges such as the difficulty in distinguishing the features of both hands, misalignment of hand meshes with input images, and modeling the complex spatial relationships between interacting hands. In this paper, we propose a multilevel feature fusion interactive network for hand reconstruction (HandFI). Within this network, the hand feature separation module utilizes attentional mechanisms and positional coding to distinguish between left-hand and right-hand features while maintaining the spatial relationship of the features. The hand fusion and attention module promotes the alignment of hand vertices with the image by integrating multi-scale hand features while introducing cross-attention to help determine the complex spatial relationships between interacting hands, thereby enhancing the accuracy of two-hand reconstruction. We evaluated our method with existing approaches using the InterHand 2.6M, RGB2Hands, and EgoHands datasets. Extensive experimental results demonstrated that our method outperformed other representative methods, with performance metrics of 9.38 mm for the MPJPE and 9.61 mm for the MPVPE. Additionally, the results obtained in real-world scenes further validated the generalization capability of our method.

摘要

交互手部重建在各种应用中展现出了重大机遇。然而,它目前面临着诸多挑战,比如难以区分两只手的特征、手部网格与输入图像的对齐问题,以及对交互手部之间复杂空间关系进行建模。在本文中,我们提出了一种用于手部重建的多级特征融合交互网络(HandFI)。在这个网络中,手部特征分离模块利用注意力机制和位置编码来区分左手和右手特征,同时保持特征的空间关系。手部融合与注意力模块通过整合多尺度手部特征来促进手部顶点与图像的对齐,同时引入交叉注意力以帮助确定交互手部之间的复杂空间关系,从而提高双手重建的准确性。我们使用InterHand 2.6M、RGB2Hands和EgoHands数据集,将我们的方法与现有方法进行了评估。大量实验结果表明,我们的方法优于其他代表性方法,平均关节位置误差(MPJPE)的性能指标为9.38毫米,平均顶点位置误差(MPVPE)为9.61毫米。此外,在真实场景中获得的结果进一步验证了我们方法的泛化能力。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2d5/11722860/6b49d8c4cdb9/sensors-25-00088-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2d5/11722860/831e507d47e9/sensors-25-00088-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2d5/11722860/2b0d108a7026/sensors-25-00088-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2d5/11722860/ac83fc9fec4f/sensors-25-00088-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2d5/11722860/6f605f738dae/sensors-25-00088-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2d5/11722860/8d6af39da32d/sensors-25-00088-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2d5/11722860/498d4ce9921d/sensors-25-00088-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2d5/11722860/6b49d8c4cdb9/sensors-25-00088-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2d5/11722860/831e507d47e9/sensors-25-00088-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2d5/11722860/2b0d108a7026/sensors-25-00088-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2d5/11722860/ac83fc9fec4f/sensors-25-00088-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2d5/11722860/6f605f738dae/sensors-25-00088-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2d5/11722860/8d6af39da32d/sensors-25-00088-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2d5/11722860/498d4ce9921d/sensors-25-00088-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2d5/11722860/6b49d8c4cdb9/sensors-25-00088-g007.jpg

相似文献

1
HandFI: Multilevel Interacting Hand Reconstruction Based on Multilevel Feature Fusion in RGB Images.HandFI:基于RGB图像中多级别特征融合的多级别交互手部重建
Sensors (Basel). 2024 Dec 27;25(1):88. doi: 10.3390/s25010088.
2
Hyperspectral Image Super Resolution With Real Unaligned RGB Guidance.基于真实未对齐RGB引导的高光谱图像超分辨率
IEEE Trans Neural Netw Learn Syst. 2025 Feb;36(2):2999-3011. doi: 10.1109/TNNLS.2023.3340561. Epub 2025 Feb 6.
3
A multi-scale pyramid residual weight network for medical image fusion.一种用于医学图像融合的多尺度金字塔残差权重网络。
Quant Imaging Med Surg. 2025 Mar 3;15(3):1793-1821. doi: 10.21037/qims-24-851. Epub 2025 Feb 26.
4
A Multi-Scale Recursive Attention Feature Fusion Network for Image Super-Resolution Reconstruction Algorithm.一种用于图像超分辨率重建算法的多尺度递归注意力特征融合网络
Sensors (Basel). 2023 Nov 28;23(23):9458. doi: 10.3390/s23239458.
5
Joint Hand-Object 3D Reconstruction From a Single Image With Cross-Branch Feature Fusion.基于跨分支特征融合的单目图像手-目标 3D 重建
IEEE Trans Image Process. 2021;30:4008-4021. doi: 10.1109/TIP.2021.3068645. Epub 2021 Apr 5.
6
CrossFuNet: RGB and Depth Cross-Fusion Network for Hand Pose Estimation.CrossFuNet:用于手部姿势估计的 RGB 和深度交叉融合网络。
Sensors (Basel). 2021 Sep 11;21(18):6095. doi: 10.3390/s21186095.
7
OIF-Net: An Optical Flow Registration-Based PET/MR Cross-Modal Interactive Fusion Network for Low-Count Brain PET Image Denoising.OIF-Net:一种基于光流配准的 PET/MR 跨模态交互融合网络,用于低计数脑 PET 图像去噪。
IEEE Trans Med Imaging. 2024 Apr;43(4):1554-1567. doi: 10.1109/TMI.2023.3342809. Epub 2024 Apr 3.
8
A hybrid network for fiber orientation distribution reconstruction employing multi-scale information.一种采用多尺度信息的用于纤维取向分布重建的混合网络。
Med Phys. 2025 Feb;52(2):1019-1036. doi: 10.1002/mp.17505. Epub 2024 Nov 20.
9
Transformer guided self-adaptive network for multi-scale skin lesion image segmentation.Transformer 引导的自适网络用于多尺度皮肤病变图像分割。
Comput Biol Med. 2024 Feb;169:107846. doi: 10.1016/j.compbiomed.2023.107846. Epub 2023 Dec 23.
10
An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion.一种基于多级语义融合的交互式图像分割方法。
Sensors (Basel). 2023 Jul 14;23(14):6394. doi: 10.3390/s23146394.

引用本文的文献

1
Lightweight Explicit 3D Human Digitization via Normal Integration.通过法线积分实现轻量级显式3D人体数字化
Sensors (Basel). 2025 Feb 28;25(5):1513. doi: 10.3390/s25051513.

本文引用的文献

1
Lending A Hand: Detecting Hands and Recognizing Activities in Complex Egocentric Interactions.伸出援手:在复杂的自我中心交互中检测手部动作并识别活动。
Proc IEEE Int Conf Comput Vis. 2015 Dec;2015:1949-1957. doi: 10.1109/ICCV.2015.226. Epub 2016 Feb 18.