• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于改进的MVSNet的增强多视图3D重建。

Enhanced multi view 3D reconstruction with improved MVSNet.

作者信息

Li Guangchen, Li Kefeng, Zhang Guangyuan, Zhu Zhenfang, Wang Peng, Wang Zhenfei, Fu Chen

机构信息

Shandong Jiaotong University, Haitang Road 5001, Jinan, 250357, China.

Shandong Zhengyuan Yeda Environmental Technology Co., Ltd., Jinan, 250101, China.

出版信息

Sci Rep. 2024 Jun 19;14(1):14106. doi: 10.1038/s41598-024-64805-y.

DOI:10.1038/s41598-024-64805-y
PMID:38890489
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11189500/
Abstract

Although 3D reconstruction has been widely used in many fields as a key component of environment perception, existing technologies still have the potential for further improvement in 3D scene reconstruction. We propose an improved reconstruction algorithm based on the MVSNet network architecture. To glean richer pixel details from images, we suggest deploying a DE module integrated with a residual framework, which supplants the prevailing feature extraction mechanism. The DE module uses ECA-Net and dilated convolution to expand the receptive field range, performing feature splicing and fusion through the residual structure to retain the global information of the original image. Moreover, harnessing attention mechanisms refines the 3D cost volume's regularization process, bolstering the integration of information across multi-scale feature volumes, consequently enhancing depth estimation precision. When assessed our model using the DTU dataset, findings highlight the network's 3D reconstruction scoring a completeness (comp) of 0.411 mm and an overall quality of 0.418 mm. This performance is higher than that of traditional methods and other deep learning-based methods. Additionally, the visual representation of the point cloud model exhibits marked advancements. Trials on the Blended MVS dataset signify that our network exhibits commendable generalization prowess.

摘要

尽管三维重建作为环境感知的关键组成部分已在许多领域得到广泛应用,但现有技术在三维场景重建方面仍有进一步改进的潜力。我们提出了一种基于MVSNet网络架构的改进重建算法。为了从图像中获取更丰富的像素细节,我们建议部署一个集成了残差框架的DE模块,以取代现有的特征提取机制。DE模块使用ECA-Net和空洞卷积来扩大感受野范围,通过残差结构进行特征拼接和融合,以保留原始图像的全局信息。此外,利用注意力机制优化三维代价体的正则化过程,加强跨多尺度特征体的信息整合,从而提高深度估计精度。当使用DTU数据集评估我们的模型时,结果表明该网络的三维重建完整性(comp)为0.411毫米,整体质量为0.418毫米。这一性能高于传统方法和其他基于深度学习的方法。此外,点云模型的视觉表现有显著进步。在混合MVS数据集上的试验表明,我们的网络具有出色的泛化能力。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/3617d2aba6d8/41598_2024_64805_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/c27dfa3b7e05/41598_2024_64805_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/e1da764f9bda/41598_2024_64805_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/f2899db60b7f/41598_2024_64805_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/52501f141997/41598_2024_64805_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/921ee0e8611e/41598_2024_64805_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/a7b709e4245e/41598_2024_64805_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/f724466374e4/41598_2024_64805_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/98982219aeb8/41598_2024_64805_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/29abbebce6c2/41598_2024_64805_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/76ad74e2306e/41598_2024_64805_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/3617d2aba6d8/41598_2024_64805_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/c27dfa3b7e05/41598_2024_64805_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/e1da764f9bda/41598_2024_64805_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/f2899db60b7f/41598_2024_64805_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/52501f141997/41598_2024_64805_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/921ee0e8611e/41598_2024_64805_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/a7b709e4245e/41598_2024_64805_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/f724466374e4/41598_2024_64805_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/98982219aeb8/41598_2024_64805_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/29abbebce6c2/41598_2024_64805_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/76ad74e2306e/41598_2024_64805_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/911a/11189500/3617d2aba6d8/41598_2024_64805_Fig11_HTML.jpg

相似文献

1
Enhanced multi view 3D reconstruction with improved MVSNet.基于改进的MVSNet的增强多视图3D重建。
Sci Rep. 2024 Jun 19;14(1):14106. doi: 10.1038/s41598-024-64805-y.
2
DRI-MVSNet: A depth residual inference network for multi-view stereo images.DRI-MVSNet:一种用于多视角立体图像的深度残差推理网络。
PLoS One. 2022 Mar 23;17(3):e0264721. doi: 10.1371/journal.pone.0264721. eCollection 2022.
3
OD-MVSNet: Omni-dimensional dynamic multi-view stereo network.OD-MVSNet:全维动态多视角立体网络。
PLoS One. 2024 Aug 15;19(8):e0309029. doi: 10.1371/journal.pone.0309029. eCollection 2024.
4
Visibility-Aware Point-Based Multi-View Stereo Network.基于可见性感知点的多视图立体视觉网络
IEEE Trans Pattern Anal Mach Intell. 2021 Oct;43(10):3695-3708. doi: 10.1109/TPAMI.2020.2988729. Epub 2021 Sep 2.
5
EI-MVSNet: Epipolar-Guided Multi-View Stereo Network With Interval-Aware Label.EI-MVSNet:具有区间感知标签的极线引导多视图立体视觉网络
IEEE Trans Image Process. 2024;33:753-766. doi: 10.1109/TIP.2023.3347929. Epub 2024 Jan 12.
6
A Light Multi-View Stereo Method with Patch-Uncertainty Awareness.一种具有面片不确定性感知的轻量级多视图立体方法。
Sensors (Basel). 2024 Feb 17;24(4):1293. doi: 10.3390/s24041293.
7
BSI-MVS: multi-view stereo network with bidirectional semantic information.BSI-MVS:具有双向语义信息的多视图立体网络。
Sci Rep. 2024 Mar 21;14(1):6766. doi: 10.1038/s41598-024-55612-6.
8
Miper-MVS: Multi-scale iterative probability estimation with refinement for efficient multi-view stereo.Miper-MVS:具有细化的多尺度迭代概率估计的高效多视图立体。
Neural Netw. 2023 May;162:502-515. doi: 10.1016/j.neunet.2023.03.012. Epub 2023 Mar 17.
9
NR-MVSNet: Learning Multi-View Stereo Based on Normal Consistency and Depth Refinement.NR-MVSNet:基于法向一致性和深度细化的多视图立体学习。
IEEE Trans Image Process. 2023;32:2649-2662. doi: 10.1109/TIP.2023.3272170. Epub 2023 May 12.
10
[Fully Automatic Glioma Segmentation Algorithm of Magnetic Resonance Imaging Based on 3D-UNet With More Global Contextual Feature Extraction: An Improvement on Insufficient Extraction of Global Features].基于具有更多全局上下文特征提取的3D-UNet的磁共振成像全自动胶质瘤分割算法:对全局特征提取不足的改进
Sichuan Da Xue Xue Bao Yi Xue Ban. 2024 Mar 20;55(2):447-454. doi: 10.12182/20240360208.

引用本文的文献

1
Fourier Lightfield Multiview Stereoscope for Large Field-of-View 3D Imaging in Microsurgical Settings.用于显微手术环境中大视野三维成像的傅里叶光场多视图立体显微镜
Adv Photonics Nexus. 2025 Jun;4(4). doi: 10.1117/1.apn.4.4.046008. Epub 2025 Jun 30.

本文引用的文献

1
Accurate, dense, and robust multiview stereopsis.精确、密集且鲁棒的多视图立体视觉。
IEEE Trans Pattern Anal Mach Intell. 2010 Aug;32(8):1362-76. doi: 10.1109/TPAMI.2009.161.
2
Stereo processing by semiglobal matching and mutual information.通过半全局匹配和互信息进行立体处理。
IEEE Trans Pattern Anal Mach Intell. 2008 Feb;30(2):328-41. doi: 10.1109/TPAMI.2007.1166.