• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于多视角立体视觉的多步深度增强细化网络

Multi-step depth enhancement refine network with multi-view stereo.

作者信息

Ding Yuxuan, Li Kefeng, Zhang Guangyuan, Zhu Zhenfang, Wang Peng, Wang Zhenfei, Fu Chen, Li Guangchen, Pan Ke

机构信息

College of Information Science and Electrical Engineering, Shandong Jiaotong University, Jinan, Shandong, China.

Shandong Zhengyuan Yeda Environmental Technology Co., Ltd, Jinan, Shandong, China.

出版信息

PLoS One. 2025 Feb 13;20(2):e0314418. doi: 10.1371/journal.pone.0314418. eCollection 2025.

DOI:10.1371/journal.pone.0314418
PMID:39946337
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11824967/
Abstract

This paper introduces an innovative multi-view stereo matching network-the Multi-Step Depth Enhancement Refine Network (MSDER-MVS), aimed at improving the accuracy and computational efficiency of high-resolution 3D reconstruction. The MSDER-MVS network leverages the potent capabilities of modern deep learning in conjunction with the geometric intuition of traditional 3D reconstruction techniques, with a particular focus on optimizing the quality of the depth map and the efficiency of the reconstruction process.Our key innovations include a dual-branch fusion structure and a Feature Pyramid Network (FPN) to effectively extract and integrate multi-scale features. With this approach, we construct depth maps progressively from coarse to fine, continuously improving depth prediction accuracy at each refinement stage. For cost volume construction, we employ a variance-based metric to integrate information from multiple perspectives, optimizing the consistency of the estimates. Moreover, we introduce a differentiable depth optimization process that iteratively enhances the quality of depth estimation using residuals and the Jacobian matrix, without the need for additional learnable parameters. This innovation significantly increases the network's convergence rate and the fineness of depth prediction.Extensive experiments on the standard DTU dataset (Aanas H, 2016) show that MSDER-MVS surpasses current advanced methods in accuracy, completeness, and overall performance metrics. Particularly in scenarios rich in detail, our method more precisely recovers surface details and textures, demonstrating its effectiveness and superiority for practical applications.Overall, the MSDER-MVS network offers a robust solution for precise and efficient 3D scene reconstruction. Looking forward, we aim to extend this approach to more complex environments and larger-scale datasets, further enhancing the model's generalization and real-time processing capabilities, and promoting the widespread deployment of multi-view stereo matching technology in practical applications.

摘要

本文介绍了一种创新的多视图立体匹配网络——多步深度增强细化网络(MSDER-MVS),旨在提高高分辨率三维重建的准确性和计算效率。MSDER-MVS网络利用现代深度学习的强大功能,并结合传统三维重建技术的几何直觉,特别关注优化深度图的质量和重建过程的效率。我们的关键创新包括双分支融合结构和特征金字塔网络(FPN),以有效地提取和整合多尺度特征。通过这种方法,我们从粗到精逐步构建深度图,在每个细化阶段不断提高深度预测的准确性。对于代价体构建,我们采用基于方差的度量来整合来自多个视角的信息,优化估计的一致性。此外,我们引入了一种可微的深度优化过程,该过程使用残差和雅可比矩阵迭代提高深度估计的质量,而无需额外的可学习参数。这一创新显著提高了网络的收敛速度和深度预测的精细度。在标准DTU数据集(Aanas H,2016)上进行的大量实验表明,MSDER-MVS在准确性、完整性和整体性能指标方面超过了当前的先进方法。特别是在细节丰富的场景中,我们的方法能够更精确地恢复表面细节和纹理,证明了其在实际应用中的有效性和优越性。总体而言,MSDER-MVS网络为精确高效的三维场景重建提供了一个强大的解决方案。展望未来,我们旨在将这种方法扩展到更复杂的环境和更大规模的数据集,进一步提高模型的泛化能力和实时处理能力,并推动多视图立体匹配技术在实际应用中的广泛部署。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6882/11824967/4554f827bf6e/pone.0314418.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6882/11824967/373576adca5e/pone.0314418.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6882/11824967/c46fb857c449/pone.0314418.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6882/11824967/bfbae21c2507/pone.0314418.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6882/11824967/98902783d839/pone.0314418.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6882/11824967/603dd01c0e66/pone.0314418.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6882/11824967/4554f827bf6e/pone.0314418.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6882/11824967/373576adca5e/pone.0314418.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6882/11824967/c46fb857c449/pone.0314418.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6882/11824967/bfbae21c2507/pone.0314418.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6882/11824967/98902783d839/pone.0314418.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6882/11824967/603dd01c0e66/pone.0314418.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6882/11824967/4554f827bf6e/pone.0314418.g006.jpg

相似文献

1
Multi-step depth enhancement refine network with multi-view stereo.基于多视角立体视觉的多步深度增强细化网络
PLoS One. 2025 Feb 13;20(2):e0314418. doi: 10.1371/journal.pone.0314418. eCollection 2025.
2
LNMVSNet: A Low-Noise Multi-View Stereo Depth Inference Method for 3D Reconstruction.LNMVSNet:一种用于三维重建的低噪声多视图立体深度推理方法
Sensors (Basel). 2024 Apr 9;24(8):2400. doi: 10.3390/s24082400.
3
OD-MVSNet: Omni-dimensional dynamic multi-view stereo network.OD-MVSNet:全维动态多视角立体网络。
PLoS One. 2024 Aug 15;19(8):e0309029. doi: 10.1371/journal.pone.0309029. eCollection 2024.
4
Brain tumor segmentation and detection in MRI using convolutional neural networks and VGG16.使用卷积神经网络和VGG16在磁共振成像(MRI)中进行脑肿瘤分割与检测
Cancer Biomark. 2025 Mar;42(3):18758592241311184. doi: 10.1177/18758592241311184. Epub 2025 Apr 4.
5
BSI-MVS: multi-view stereo network with bidirectional semantic information.BSI-MVS:具有双向语义信息的多视图立体网络。
Sci Rep. 2024 Mar 21;14(1):6766. doi: 10.1038/s41598-024-55612-6.
6
A Light Multi-View Stereo Method with Patch-Uncertainty Awareness.一种具有面片不确定性感知的轻量级多视图立体方法。
Sensors (Basel). 2024 Feb 17;24(4):1293. doi: 10.3390/s24041293.
7
Neural Radiance Field-Inspired Depth Map Refinement for Accurate Multi-View Stereo.受神经辐射场启发的深度图细化用于精确多视图立体视觉
J Imaging. 2024 Mar 8;10(3):68. doi: 10.3390/jimaging10030068.
8
Miper-MVS: Multi-scale iterative probability estimation with refinement for efficient multi-view stereo.Miper-MVS:具有细化的多尺度迭代概率估计的高效多视图立体。
Neural Netw. 2023 May;162:502-515. doi: 10.1016/j.neunet.2023.03.012. Epub 2023 Mar 17.
9
Visibility-Aware Point-Based Multi-View Stereo Network.基于可见性感知点的多视图立体视觉网络
IEEE Trans Pattern Anal Mach Intell. 2021 Oct;43(10):3695-3708. doi: 10.1109/TPAMI.2020.2988729. Epub 2021 Sep 2.
10
Optimization of sparse-view CT reconstruction based on convolutional neural network.基于卷积神经网络的稀疏视图CT重建优化
Med Phys. 2025 Apr;52(4):2089-2105. doi: 10.1002/mp.17636. Epub 2025 Feb 2.

引用本文的文献

1
Fourier Lightfield Multiview Stereoscope for Large Field-of-View 3D Imaging in Microsurgical Settings.用于显微手术环境中大视野三维成像的傅里叶光场多视图立体显微镜
Adv Photonics Nexus. 2025 Jun;4(4). doi: 10.1117/1.apn.4.4.046008. Epub 2025 Jun 30.

本文引用的文献

1
Accurate, dense, and robust multiview stereopsis.精确、密集且鲁棒的多视图立体视觉。
IEEE Trans Pattern Anal Mach Intell. 2010 Aug;32(8):1362-76. doi: 10.1109/TPAMI.2009.161.