• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

DRI-MVSNet:一种用于多视角立体图像的深度残差推理网络。

DRI-MVSNet: A depth residual inference network for multi-view stereo images.

机构信息

College of Computer Science and Technology, Jilin University, Changchun, China.

Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, Jilin University, Changchun, China.

出版信息

PLoS One. 2022 Mar 23;17(3):e0264721. doi: 10.1371/journal.pone.0264721. eCollection 2022.

DOI:10.1371/journal.pone.0264721
PMID:35320265
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8942269/
Abstract

Three-dimensional (3D) image reconstruction is an important field of computer vision for restoring the 3D geometry of a given scene. Due to the demand for large amounts of memory, prevalent methods of 3D reconstruction yield inaccurate results, because of which the highly accuracy reconstruction of a scene remains an outstanding challenge. This study proposes a cascaded depth residual inference network, called DRI-MVSNet, that uses a cross-view similarity-based feature map fusion module for residual inference. It involves three improvements. First, a combined module is used for processing channel-related and spatial information to capture the relevant contextual information and improve feature representation. It combines the channel attention mechanism and spatial pooling networks. Second, a cross-view similarity-based feature map fusion module is proposed that learns the similarity between pairs of pixel in each source and reference image at planes of different depths along the frustum of the reference camera. Third, a deep, multi-stage residual prediction module is designed to generate a high-precision depth map that uses a non-uniform depth sampling strategy to construct hypothetical depth planes. The results of extensive experiments show that DRI-MVSNet delivers competitive performance on the DTU and the Tanks & Temples datasets, and the accuracy and completeness of the point cloud reconstructed by it are significantly superior to those of state-of-the-art benchmarks.

摘要

三维(3D)图像重建是计算机视觉领域的一个重要分支,用于恢复给定场景的 3D 几何形状。由于需要大量的内存,目前主流的 3D 重建方法的结果不够准确,因此场景的高精度重建仍然是一个具有挑战性的问题。本研究提出了一种级联深度残差推理网络,称为 DRI-MVSNet,它使用基于跨视图相似性的特征图融合模块进行残差推理。该方法有三个改进点。首先,使用了一个组合模块来处理与通道相关和空间信息,以捕获相关的上下文信息并改进特征表示。它结合了通道注意力机制和空间池化网络。其次,提出了一种基于跨视图相似性的特征图融合模块,用于学习参考相机光锥中不同深度平面上每对源和参考图像中像素之间的相似性。第三,设计了一个深度、多阶段的残差预测模块,用于生成高精度的深度图。该模块使用非均匀深度采样策略来构建假设深度平面。在广泛的实验结果表明,DRI-MVSNet 在 DTU 和 Tanks & Temples 数据集上表现出了有竞争力的性能,并且它重建的点云的准确性和完整性明显优于最新的基准。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e862/8942269/13d9b1545410/pone.0264721.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e862/8942269/a95f3f3d565c/pone.0264721.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e862/8942269/a71fbe0896d7/pone.0264721.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e862/8942269/5090a0a4d559/pone.0264721.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e862/8942269/00e87ae9ac4d/pone.0264721.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e862/8942269/46ca3b81987f/pone.0264721.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e862/8942269/13d9b1545410/pone.0264721.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e862/8942269/a95f3f3d565c/pone.0264721.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e862/8942269/a71fbe0896d7/pone.0264721.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e862/8942269/5090a0a4d559/pone.0264721.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e862/8942269/00e87ae9ac4d/pone.0264721.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e862/8942269/46ca3b81987f/pone.0264721.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e862/8942269/13d9b1545410/pone.0264721.g006.jpg

相似文献

1
DRI-MVSNet: A depth residual inference network for multi-view stereo images.DRI-MVSNet:一种用于多视角立体图像的深度残差推理网络。
PLoS One. 2022 Mar 23;17(3):e0264721. doi: 10.1371/journal.pone.0264721. eCollection 2022.
2
OD-MVSNet: Omni-dimensional dynamic multi-view stereo network.OD-MVSNet:全维动态多视角立体网络。
PLoS One. 2024 Aug 15;19(8):e0309029. doi: 10.1371/journal.pone.0309029. eCollection 2024.
3
NR-MVSNet: Learning Multi-View Stereo Based on Normal Consistency and Depth Refinement.NR-MVSNet:基于法向一致性和深度细化的多视图立体学习。
IEEE Trans Image Process. 2023;32:2649-2662. doi: 10.1109/TIP.2023.3272170. Epub 2023 May 12.
4
Enhanced multi view 3D reconstruction with improved MVSNet.基于改进的MVSNet的增强多视图3D重建。
Sci Rep. 2024 Jun 19;14(1):14106. doi: 10.1038/s41598-024-64805-y.
5
Visibility-Aware Point-Based Multi-View Stereo Network.基于可见性感知点的多视图立体视觉网络
IEEE Trans Pattern Anal Mach Intell. 2021 Oct;43(10):3695-3708. doi: 10.1109/TPAMI.2020.2988729. Epub 2021 Sep 2.
6
EI-MVSNet: Epipolar-Guided Multi-View Stereo Network With Interval-Aware Label.EI-MVSNet:具有区间感知标签的极线引导多视图立体视觉网络
IEEE Trans Image Process. 2024;33:753-766. doi: 10.1109/TIP.2023.3347929. Epub 2024 Jan 12.
7
Miper-MVS: Multi-scale iterative probability estimation with refinement for efficient multi-view stereo.Miper-MVS:具有细化的多尺度迭代概率估计的高效多视图立体。
Neural Netw. 2023 May;162:502-515. doi: 10.1016/j.neunet.2023.03.012. Epub 2023 Mar 17.
8
A Light Multi-View Stereo Method with Patch-Uncertainty Awareness.一种具有面片不确定性感知的轻量级多视图立体方法。
Sensors (Basel). 2024 Feb 17;24(4):1293. doi: 10.3390/s24041293.
9
BSI-MVS: multi-view stereo network with bidirectional semantic information.BSI-MVS:具有双向语义信息的多视图立体网络。
Sci Rep. 2024 Mar 21;14(1):6766. doi: 10.1038/s41598-024-55612-6.
10
RayMVSNet++: Learning Ray-Based 1D Implicit Fields for Accurate Multi-View Stereo.RayMVSNet++:学习基于光线的一维隐式场以实现精确的多视图立体视觉。
IEEE Trans Pattern Anal Mach Intell. 2023 Nov;45(11):13666-13682. doi: 10.1109/TPAMI.2023.3296163. Epub 2023 Oct 3.

本文引用的文献

1
Cost Volume Pyramid Based Depth Inference for Multi-View Stereo.基于成本体积金字塔的多视图立体深度推理
IEEE Trans Pattern Anal Mach Intell. 2022 Sep;44(9):4748-4760. doi: 10.1109/TPAMI.2021.3082562. Epub 2022 Aug 4.
2
SurfaceNet+: An End-to-end 3D Neural Network for Very Sparse Multi-View Stereopsis.SurfaceNet+:用于极稀疏多视图立体视觉的端到端3D神经网络。
IEEE Trans Pattern Anal Mach Intell. 2021 Nov;43(11):4078-4093. doi: 10.1109/TPAMI.2020.2996798. Epub 2021 Oct 1.
3
Visibility-Aware Point-Based Multi-View Stereo Network.
基于可见性感知点的多视图立体视觉网络
IEEE Trans Pattern Anal Mach Intell. 2021 Oct;43(10):3695-3708. doi: 10.1109/TPAMI.2020.2988729. Epub 2021 Sep 2.
4
Autonomous Vehicles: Disengagements, Accidents and Reaction Times.自动驾驶汽车:脱离、事故与反应时间。
PLoS One. 2016 Dec 20;11(12):e0168054. doi: 10.1371/journal.pone.0168054. eCollection 2016.
5
Accurate multiple view 3D reconstruction using patch-based stereo for large-scale scenes.基于补丁的立体视觉的精确多视图 3D 重建用于大规模场景。
IEEE Trans Image Process. 2013 May;22(5):1901-14. doi: 10.1109/TIP.2013.2237921. Epub 2013 Jan 10.
6
Bistable percepts in the brain: FMRI contrasts monocular pattern rivalry and binocular rivalry.大脑中的双稳态知觉:FMRI 对比单眼模式竞争和双眼竞争。
PLoS One. 2011;6(5):e20367. doi: 10.1371/journal.pone.0020367. Epub 2011 May 23.
7
Addressing overutilization in medical imaging.解决医学影像中的过度使用问题。
Radiology. 2010 Oct;257(1):240-5. doi: 10.1148/radiol.10100063. Epub 2010 Aug 24.
8
Accurate, dense, and robust multiview stereopsis.精确、密集且鲁棒的多视图立体视觉。
IEEE Trans Pattern Anal Mach Intell. 2010 Aug;32(8):1362-76. doi: 10.1109/TPAMI.2009.161.