• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

双通道引导聚合网络的立体图像匹配。

Dual Guided Aggregation Network for Stereo Image Matching.

机构信息

Department of Geomatics, National Cheng-Kung University, Tainan City 701, Taiwan.

出版信息

Sensors (Basel). 2022 Aug 16;22(16):6111. doi: 10.3390/s22166111.

DOI:10.3390/s22166111
PMID:36015872
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9414513/
Abstract

Stereo image dense matching, which plays a key role in 3D reconstruction, remains a challenging task in photogrammetry and computer vision. In addition to block-based matching, recent studies based on artificial neural networks have achieved great progress in stereo matching by using deep convolutional networks. This study proposes a novel network called a dual guided aggregation network (Dual-GANet), which utilizes both left-to-right and right-to-left image matchings in network design and training to reduce the possibility of pixel mismatch. Flipped training with a cost volume consistentization is introduced to realize the learning of invisible-to-visible pixel matching and left−right consistency matching. In addition, suppressed multi-regression is proposed, which suppresses unrelated information before regression and selects multiple peaks from a disparity probability distribution. The proposed dual network with the left−right consistent matching scheme can be applied to most stereo matching models. To estimate the performance, GANet, which is designed based on semi-global matching, was selected as the backbone with extensions and modifications on guided aggregation, disparity regression, and loss function. Experimental results on the SceneFlow and KITTI2015 datasets demonstrate the superiority of the Dual-GANet compared to related models in terms of average end-point-error (EPE) and pixel error rate (ER). The Dual-GANet with an average EPE performance = 0.418 and ER (>1 pixel) = 5.81% for SceneFlow and average EPE = 0.589 and ER (>3 pixels) = 1.76% for KITTI2005 is better than the backbone model with the average EPE performance of = 0.440 and ER (>1 pixel) = 6.56% for SceneFlow and average EPE = 0.790 and ER (>3 pixels) = 2.32% for KITTI2005.

摘要

立体图像密集匹配在 3D 重建中起着关键作用,仍然是摄影测量和计算机视觉中的一项具有挑战性的任务。除了基于块的匹配外,基于人工神经网络的最新研究通过使用深度卷积网络在立体匹配方面取得了很大的进展。本研究提出了一种名为双引导聚合网络(Dual-GANet)的新网络,该网络在网络设计和训练中利用左右图像匹配来减少像素不匹配的可能性。引入翻转训练和代价体一致性化,以实现不可见-可见像素匹配和左右一致性匹配的学习。此外,提出了抑制多回归,在回归之前抑制不相关的信息,并从视差概率分布中选择多个峰值。所提出的具有左右一致匹配方案的双网络可以应用于大多数立体匹配模型。为了评估性能,选择基于半全局匹配设计的 GANet 作为骨干,并在引导聚合、视差回归和损失函数方面进行扩展和修改。在 SceneFlow 和 KITTI2015 数据集上的实验结果表明,与相关模型相比,Dual-GANet 在平均端点误差(EPE)和像素误差率(ER)方面具有优越性。对于 SceneFlow,Dual-GANet 的平均 EPE 性能为 0.418,ER (>1 像素) 为 5.81%,ER (>3 像素) 为 1.76%;对于 KITTI2005,Dual-GANet 的平均 EPE 性能为 0.589,ER (>1 像素) 为 6.56%,ER (>3 像素) 为 1.76%。优于骨干模型,对于 SceneFlow,其平均 EPE 性能为 0.440,ER (>1 像素) 为 6.56%,ER (>3 像素) 为 2.32%;对于 KITTI2005,其平均 EPE 性能为 0.790,ER (>1 像素) 为 2.32%,ER (>3 像素) 为 2.32%。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/f8d523c58682/sensors-22-06111-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/02553f69ac77/sensors-22-06111-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/585d9f5e5e2f/sensors-22-06111-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/e1c2aa1048dc/sensors-22-06111-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/d71e9660d621/sensors-22-06111-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/018a78d42968/sensors-22-06111-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/91d9de6f8906/sensors-22-06111-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/b66d0b0a1db4/sensors-22-06111-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/d1b5fd9ec067/sensors-22-06111-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/f8d523c58682/sensors-22-06111-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/02553f69ac77/sensors-22-06111-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/585d9f5e5e2f/sensors-22-06111-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/e1c2aa1048dc/sensors-22-06111-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/d71e9660d621/sensors-22-06111-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/018a78d42968/sensors-22-06111-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/91d9de6f8906/sensors-22-06111-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/b66d0b0a1db4/sensors-22-06111-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/d1b5fd9ec067/sensors-22-06111-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/63b7/9414513/f8d523c58682/sensors-22-06111-g009.jpg

相似文献

1
Dual Guided Aggregation Network for Stereo Image Matching.双通道引导聚合网络的立体图像匹配。
Sensors (Basel). 2022 Aug 16;22(16):6111. doi: 10.3390/s22166111.
2
Cost aggregation and occlusion handling with WLS in stereo matching.立体匹配中使用加权最小二乘法(WLS)的代价聚合与遮挡处理
IEEE Trans Image Process. 2008 Aug;17(8):1431-42. doi: 10.1109/TIP.2008.925372.
3
A study of stereo microscope measurements based on interpolated feature matching.基于插值特征匹配的体视显微镜测量研究
Biomed Mater Eng. 2015;26 Suppl 1:S1473-81. doi: 10.3233/BME-151446.
4
Joint histogram-based cost aggregation for stereo matching.基于联合直方图的立体匹配代价聚合。
IEEE Trans Pattern Anal Mach Intell. 2013 Oct;35(10):2539-45. doi: 10.1109/TPAMI.2013.15.
5
Discriminability limits in spatio-temporal stereo block matching.时空间隔立体块匹配中的可辨别性限制。
IEEE Trans Image Process. 2014 May;23(5):2328-42. doi: 10.1109/TIP.2014.2316420.
6
Efficient hybrid tree-based stereo matching with applications to postcapture image refocusing.基于高效混合树的立体匹配及其在后期重聚焦图像中的应用。
IEEE Trans Image Process. 2014 Aug;23(8):3428-42. doi: 10.1109/TIP.2014.2329389. Epub 2014 Jun 5.
7
On the application of a modified self-organizing neural network to estimate stereo disparity.关于应用改进的自组织神经网络来估计立体视差。
IEEE Trans Image Process. 2007 Nov;16(11):2822-9. doi: 10.1109/tip.2007.906772.
8
Fast unambiguous stereo matching using reliability-based dynamic programming.使用基于可靠性的动态规划实现快速无歧义立体匹配。
IEEE Trans Pattern Anal Mach Intell. 2005 Jun;27(6):998-1003. doi: 10.1109/TPAMI.2005.120.
9
A quantitative evaluation of confidence measures for stereo vision.立体视觉置信度度量的定量评估。
IEEE Trans Pattern Anal Mach Intell. 2012 Nov;34(11):2121-33. doi: 10.1109/TPAMI.2012.46.
10
Stereo matching with Mumford-Shah regularization and occlusion handling.基于 Mumford-Shah 正则化和遮挡处理的立体匹配。
IEEE Trans Pattern Anal Mach Intell. 2010 Nov;32(11):2071-84. doi: 10.1109/TPAMI.2010.32.

本文引用的文献

1
Enhanced Soft 3D Reconstruction Method with an Iterative Matching Cost Update Using Object Surface Consensus.基于物体表面一致性的迭代匹配代价更新的增强型软三维重建方法
Sensors (Basel). 2021 Oct 8;21(19):6680. doi: 10.3390/s21196680.
2
Learning Depth with Convolutional Spatial Propagation Network.基于卷积空间传播网络的深度学习
IEEE Trans Pattern Anal Mach Intell. 2020 Oct;42(10):2361-2379. doi: 10.1109/TPAMI.2019.2947374. Epub 2019 Oct 15.
3
Hierarchical Stereo Matching in Two-Scale Space for Cyber-Physical System.用于网络物理系统的双尺度空间分层立体匹配
Sensors (Basel). 2017 Jul 21;17(7):1680. doi: 10.3390/s17071680.
4
Stereo processing by semiglobal matching and mutual information.通过半全局匹配和互信息进行立体处理。
IEEE Trans Pattern Anal Mach Intell. 2008 Feb;30(2):328-41. doi: 10.1109/TPAMI.2007.1166.