• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

使用单个RGBD相机进行3D实时人体重建。

3D real-time human reconstruction with a single RGBD camera.

作者信息

Lu Yang, Yu Han, Ni Wei, Song Liang

机构信息

Academy of Engineering and Technology, Fudan University, Shanghai, China.

Shanghai Key Research Laboratory of NSAI, Shanghai, China.

出版信息

Appl Intell (Dordr). 2023;53(8):8735-8745. doi: 10.1007/s10489-022-03969-4. Epub 2022 Aug 2.

DOI:10.1007/s10489-022-03969-4
PMID:35937202
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9343569/
Abstract

3D human reconstruction is an important technology connecting the real world and the virtual world, but most of previous work needs expensive computing resources, making it difficult in real-time scenarios. We propose a lightweight human body reconstruction system based on parametric model, which employs only one RGBD camera as input. To generate a human model end to end, we build a fast and lightweight deep-learning network named Fast Body Net (FBN). The network pays more attention on the face and hands to enrich the local details. Additionally, we train a denoising auto-encoder to reduce unreasonable states of human model. Due to the lack of human dataset based on RGBD images, we propose an Indoor-Human dataset to train the network, which contains a total of 2500 frames of action data of five actors collected by Azure Kinect camera. Depth images avoid using RGB to extract depth features, which makes FBN lightweight and high-speed in reconstructing parametric human model. Qualitative and quantitative analysis on experimental results show that our method can improve at least 57% in efficiency with similar accuracy, as compared to state-of-the-art methods. Through our study, it is also demonstrated that consumer-grade RGBD cameras can provide great applications in real-time display and interaction for virtual reality.

摘要

三维人体重建是连接现实世界和虚拟世界的一项重要技术,但以往的大多数工作都需要昂贵的计算资源,这使得其在实时场景中应用困难。我们提出了一种基于参数模型的轻量级人体重建系统,该系统仅使用一个RGB-D相机作为输入。为了端到端地生成人体模型,我们构建了一个名为快速人体网络(FBN)的快速且轻量级的深度学习网络。该网络更加关注面部和手部以丰富局部细节。此外,我们训练了一个去噪自动编码器以减少人体模型的不合理状态。由于缺乏基于RGB-D图像的人体数据集,我们提出了一个室内人体数据集来训练该网络,它包含由Azure Kinect相机收集的五名演员的总共2500帧动作数据。深度图像避免使用RGB来提取深度特征,这使得FBN在重建参数化人体模型时既轻量级又高速。对实验结果的定性和定量分析表明,与现有方法相比,我们的方法在精度相似的情况下效率可提高至少57%。通过我们的研究还表明,消费级RGB-D相机可以在虚拟现实的实时显示和交互中提供出色的应用。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/3db0f74f0717/10489_2022_3969_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/a9a8fa4e7e6e/10489_2022_3969_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/1bdbb7677cc0/10489_2022_3969_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/6922b1aa2fa5/10489_2022_3969_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/40ce4012e34c/10489_2022_3969_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/e52e1220a1fb/10489_2022_3969_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/b69042e95f3e/10489_2022_3969_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/ea2282c312f9/10489_2022_3969_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/b2f909d0ae37/10489_2022_3969_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/3db0f74f0717/10489_2022_3969_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/a9a8fa4e7e6e/10489_2022_3969_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/1bdbb7677cc0/10489_2022_3969_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/6922b1aa2fa5/10489_2022_3969_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/40ce4012e34c/10489_2022_3969_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/e52e1220a1fb/10489_2022_3969_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/b69042e95f3e/10489_2022_3969_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/ea2282c312f9/10489_2022_3969_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/b2f909d0ae37/10489_2022_3969_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0825/9343569/3db0f74f0717/10489_2022_3969_Fig9_HTML.jpg

相似文献

1
3D real-time human reconstruction with a single RGBD camera.使用单个RGBD相机进行3D实时人体重建。
Appl Intell (Dordr). 2023;53(8):8735-8745. doi: 10.1007/s10489-022-03969-4. Epub 2022 Aug 2.
2
3D object detection through fog and occlusion: passive integral imaging vs active (LiDAR) sensing.通过雾和遮挡进行3D目标检测:被动积分成像与主动(激光雷达)传感
Opt Express. 2023 Jan 2;31(1):479-491. doi: 10.1364/OE.478125.
3
Indoor 3D Reconstruction of Buildings via Azure Kinect RGB-D Camera.基于 Azure Kinect RGB-D 相机的建筑物室内三维重建。
Sensors (Basel). 2022 Nov 27;22(23):9222. doi: 10.3390/s22239222.
4
Calibration of RGBD camera and cone-beam CT for 3D intra-operative mixed reality visualization.用于三维术中混合现实可视化的RGBD相机与锥束CT的校准
Int J Comput Assist Radiol Surg. 2016 Jun;11(6):967-75. doi: 10.1007/s11548-016-1396-1. Epub 2016 Apr 8.
5
Three-D Wide Faces (3DWF): Facial Landmark Detection and 3D Reconstruction over a New RGB⁻D Multi-Camera Dataset.三维宽面(3DWF):基于新 RGB⁻D 多摄像机数据集的面部地标检测和 3D 重建。
Sensors (Basel). 2019 Mar 4;19(5):1103. doi: 10.3390/s19051103.
6
Real-Time 3D Reconstruction Method Based on Monocular Vision.基于单目视觉的实时 3D 重建方法。
Sensors (Basel). 2021 Sep 2;21(17):5909. doi: 10.3390/s21175909.
7
A Range-Independent Disparity-Based Calibration Model for Structured Light Pattern-Based RGBD Sensor.一种基于视差的结构光 RGBD 传感器无距离依赖标定模型。
Sensors (Basel). 2020 Jan 23;20(3):639. doi: 10.3390/s20030639.
8
Real-Time Human Action Recognition with a Low-Cost RGB Camera and Mobile Robot Platform.基于低成本 RGB 相机和移动机器人平台的实时人体动作识别。
Sensors (Basel). 2020 May 19;20(10):2886. doi: 10.3390/s20102886.
9
Deep Attention Models for Human Tracking Using RGBD.基于 RGBD 的深度注意模型的人体跟踪
Sensors (Basel). 2019 Feb 13;19(4):750. doi: 10.3390/s19040750.
10
Exploring RGB+Depth Fusion for Real-Time Object Detection.探索 RGB+深度融合的实时目标检测。
Sensors (Basel). 2019 Feb 19;19(4):866. doi: 10.3390/s19040866.

引用本文的文献

1
Light-Adaptive Human Body Key Point Detection Algorithm Based on Multi-Source Information Fusion.基于多源信息融合的光适应人体关键点检测算法。
Sensors (Basel). 2024 May 10;24(10):3021. doi: 10.3390/s24103021.

本文引用的文献

1
OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields.OpenPose:基于部件亲和力字段的实时多人 2D 姿态估计。
IEEE Trans Pattern Anal Mach Intell. 2021 Jan;43(1):172-186. doi: 10.1109/TPAMI.2019.2929257. Epub 2020 Dec 4.
2
UnstructuredFusion: Realtime 4D Geometry and Texture Reconstruction Using Commercial RGBD Cameras.非结构化融合:使用商业 RGBD 相机进行实时 4D 几何和纹理重建。
IEEE Trans Pattern Anal Mach Intell. 2020 Oct;42(10):2508-2522. doi: 10.1109/TPAMI.2019.2915229. Epub 2019 May 7.
3
Compressing VR: Fitting Large Virtual Environments within Limited Physical Space.
压缩虚拟现实:在有限物理空间内适配大型虚拟环境。
IEEE Comput Graph Appl. 2017;37(5):85-91. doi: 10.1109/MCG.2017.3621226.