• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于自适应代价体滤波的高效多尺度立体匹配网络

Efficient Multi-Scale Stereo-Matching Network Using Adaptive Cost Volume Filtering.

作者信息

Jeon Suyeon, Heo Yong Seok

机构信息

Department of Artificial Intelligence, Ajou University, Suwon 16499, Korea.

Department of Electrical and Computer Engineering, Ajou University, Suwon 16499, Korea.

出版信息

Sensors (Basel). 2022 Jul 23;22(15):5500. doi: 10.3390/s22155500.

DOI:10.3390/s22155500
PMID:35898003
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9371070/
Abstract

While recent deep learning-based stereo-matching networks have shown outstanding advances, there are still some unsolved challenges. First, most state-of-the-art stereo models employ 3D convolutions for 4D cost volume aggregation, which limit the deployment of networks for resource-limited mobile environments owing to heavy consumption of computation and memory. Although there are some efficient networks, most of them still require a heavy computational cost to incorporate them to mobile computing devices in real-time. Second, most stereo networks indirectly supervise cost volumes through disparity regression loss by using the softargmax function. This causes problems in ambiguous regions, such as the boundaries of objects, because there are many possibilities for unreasonable cost distributions which result in overfitting problem. A few works deal with this problem by generating artificial cost distribution using only the ground truth disparity value that is insufficient to fully regularize the cost volume. To address these problems, we first propose an efficient multi-scale sequential feature fusion network (MSFFNet). Specifically, we connect multi-scale SFF modules in parallel with a cross-scale fusion function to generate a set of cost volumes with different scales. These cost volumes are then effectively combined using the proposed interlaced concatenation method. Second, we propose an adaptive cost-volume-filtering (ACVF) loss function that directly supervises our estimated cost volume. The proposed ACVF loss directly adds constraints to the cost volume using the probability distribution generated from the ground truth disparity map and that estimated from the teacher network which achieves higher accuracy. Results of several experiments using representative datasets for stereo matching show that our proposed method is more efficient than previous methods. Our network architecture consumes fewer parameters and generates reasonable disparity maps with faster speed compared with the existing state-of-the art stereo models. Concretely, our network achieves 1.01 EPE with runtime of 42 ms, 2.92M parameters, and 97.96G FLOPs on the Scene Flow test set. Compared with PSMNet, our method is 89% faster and 7% more accurate with 45% fewer parameters.

摘要

尽管最近基于深度学习的立体匹配网络取得了显著进展,但仍存在一些未解决的挑战。首先,大多数最先进的立体模型采用3D卷积进行4D代价体聚合,由于计算和内存消耗大,这限制了网络在资源有限的移动环境中的部署。虽然有一些高效的网络,但大多数仍需要大量计算成本才能实时集成到移动计算设备中。其次,大多数立体网络通过使用softargmax函数通过视差回归损失间接监督代价体。这在模糊区域(如物体边界)会导致问题,因为存在许多不合理的代价分布可能性,从而导致过拟合问题。一些工作通过仅使用地面真值视差值生成人工代价分布来处理这个问题,而这不足以完全正则化代价体。为了解决这些问题,我们首先提出了一种高效的多尺度顺序特征融合网络(MSFFNet)。具体来说,我们将多尺度SFF模块与跨尺度融合函数并行连接,以生成一组不同尺度的代价体。然后使用所提出的交错连接方法有效地组合这些代价体。其次,我们提出了一种自适应代价体滤波(ACVF)损失函数,直接监督我们估计的代价体。所提出的ACVF损失使用从地面真值视差图生成的概率分布和从教师网络估计的概率分布直接向代价体添加约束,教师网络具有更高的准确性。使用代表性立体匹配数据集进行的几个实验结果表明,我们提出的方法比以前的方法更有效。与现有的最先进立体模型相比,我们的网络架构消耗的参数更少,生成合理视差图的速度更快。具体而言,我们的网络在场景流测试集上实现了1.01的端点误差(EPE),运行时间为42毫秒,有292万个参数和979.6亿次浮点运算(FLOPs)。与PSMNet相比,我们的方法速度快89%,准确率高7%,参数少45%。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/126d66c42b3a/sensors-22-05500-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/a3371506968c/sensors-22-05500-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/a45b464912fe/sensors-22-05500-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/532fb827de06/sensors-22-05500-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/9bb4cbc5a6a7/sensors-22-05500-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/9375765875db/sensors-22-05500-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/9e2fb6ace944/sensors-22-05500-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/60558dc968df/sensors-22-05500-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/6c6f7181b75a/sensors-22-05500-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/1d80ac680889/sensors-22-05500-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/999e2309a159/sensors-22-05500-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/126d66c42b3a/sensors-22-05500-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/a3371506968c/sensors-22-05500-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/a45b464912fe/sensors-22-05500-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/532fb827de06/sensors-22-05500-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/9bb4cbc5a6a7/sensors-22-05500-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/9375765875db/sensors-22-05500-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/9e2fb6ace944/sensors-22-05500-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/60558dc968df/sensors-22-05500-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/6c6f7181b75a/sensors-22-05500-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/1d80ac680889/sensors-22-05500-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/999e2309a159/sensors-22-05500-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/30cc/9371070/126d66c42b3a/sensors-22-05500-g011.jpg

相似文献

1
Efficient Multi-Scale Stereo-Matching Network Using Adaptive Cost Volume Filtering.基于自适应代价体滤波的高效多尺度立体匹配网络
Sensors (Basel). 2022 Jul 23;22(15):5500. doi: 10.3390/s22155500.
2
A stereo matching algorithm based on the improved PSMNet.基于改进的 PSMNet 的立体匹配算法。
PLoS One. 2021 Aug 19;16(8):e0251657. doi: 10.1371/journal.pone.0251657. eCollection 2021.
3
Single-shot 3D shape measurement using an end-to-end stereo matching network for speckle projection profilometry.使用端到端立体匹配网络进行散斑投影轮廓测量的单次3D形状测量。
Opt Express. 2021 Apr 26;29(9):13388-13407. doi: 10.1364/OE.418881.
4
RLStereo: Real-Time Stereo Matching Based on Reinforcement Learning.
IEEE Trans Image Process. 2021;30:9442-9455. doi: 10.1109/TIP.2021.3126418. Epub 2021 Nov 18.
5
Optimizing 3D Convolution Kernels on Stereo Matching for Resource Efficient Computations.优化立体匹配中的 3D 卷积核以实现资源高效计算。
Sensors (Basel). 2021 Oct 13;21(20):6808. doi: 10.3390/s21206808.
6
Stereo Matching Using Multi-Level Cost Volume and Multi-Scale Feature Constancy.基于多级代价体和多尺度特征一致性的立体匹配
IEEE Trans Pattern Anal Mach Intell. 2021 Jan;43(1):300-315. doi: 10.1109/TPAMI.2019.2928550. Epub 2020 Dec 4.
7
A Joint 2D-3D Complementary Network for Stereo Matching.一种用于立体匹配的二维-三维互补网络。
Sensors (Basel). 2021 Feb 18;21(4):1430. doi: 10.3390/s21041430.
8
Unified Confidence Estimation Networks for Robust Stereo Matching.用于稳健立体匹配的统一置信估计网络。
IEEE Trans Image Process. 2019 Mar;28(3):1299-1313. doi: 10.1109/TIP.2018.2878325. Epub 2018 Oct 26.
9
Parallax attention stereo matching network based on the improved group-wise correlation stereo network.基于改进的分组相关立体网络的视差注意力立体匹配网络。
PLoS One. 2022 Feb 9;17(2):e0263735. doi: 10.1371/journal.pone.0263735. eCollection 2022.
10
Accurate and Efficient Stereo Matching via Attention Concatenation Volume.通过注意力拼接体实现准确高效的立体匹配
IEEE Trans Pattern Anal Mach Intell. 2024 Apr;46(4):2461-2474. doi: 10.1109/TPAMI.2023.3335480. Epub 2024 Mar 6.

引用本文的文献

1
An application of stereo matching algorithm based on transfer learning on robots in multiple scenes.基于迁移学习的立体匹配算法在多场景机器人中的应用。
Sci Rep. 2023 Aug 6;13(1):12739. doi: 10.1038/s41598-023-39964-z.
2
Robust Estimation and Optimized Transmission of 3D Feature Points for Computer Vision on Mobile Communication Network.移动通讯网络上计算机视觉的三维特征点稳健估计与优化传输
Sensors (Basel). 2022 Nov 7;22(21):8563. doi: 10.3390/s22218563.