• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

光流估计的隐式和显式正则化。

Implicit and Explicit Regularization for Optical Flow Estimation.

机构信息

The Visual Computing Lab, Information Technologies Institute, Centre for Research and Technology Hellas, 57001 Thessaloniki, Greece.

Universidad Politécnica de Madrid, 28040 Madrid, Spain.

出版信息

Sensors (Basel). 2020 Jul 10;20(14):3855. doi: 10.3390/s20143855.

DOI:10.3390/s20143855
PMID:32664442
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7412516/
Abstract

In this paper, two novel and practical regularizing methods are proposed to improve existing neural network architectures for monocular optical flow estimation. The proposed methods aim to alleviate deficiencies of current methods, such as flow leakage across objects and motion consistency within rigid objects, by exploiting contextual information. More specifically, the first regularization method utilizes semantic information during the training process to explicitly regularize the produced optical flow field. The novelty of this method lies in the use of semantic segmentation masks to teach the network to implicitly identify the semantic edges of an object and better reason on the local motion flow. A novel loss function is introduced that takes into account the objects' boundaries as derived from the semantic segmentation mask to selectively penalize motion inconsistency within an object. The method is architecture agnostic and can be integrated into any neural network without modifying or adding complexity at inference. The second regularization method adds spatial awareness to the input data of the network in order to improve training stability and efficiency. The coordinates of each pixel are used as an additional feature, breaking the invariance properties of the neural network architecture. The additional features are shown to implicitly regularize the optical flow estimation enforcing a consistent flow, while improving both the performance and the convergence time. Finally, the combination of both regularization methods further improves the performance of existing cutting edge architectures in a complementary way, both quantitatively and qualitatively, on popular flow estimation benchmark datasets.

摘要

在本文中,我们提出了两种新颖而实用的正则化方法,以改进现有的用于单目光流估计的神经网络架构。所提出的方法旨在通过利用上下文信息来缓解当前方法的缺陷,例如物体之间的流泄漏和刚体内部的运动一致性。更具体地说,第一种正则化方法在训练过程中利用语义信息来显式地正则化产生的光流场。该方法的新颖之处在于使用语义分割掩模来教导网络隐式识别物体的语义边缘,并更好地推理局部运动流。引入了一种新的损失函数,该函数考虑了来自语义分割掩模的物体边界,以选择性地惩罚物体内部的运动不一致性。该方法与架构无关,可以集成到任何神经网络中,而无需在推理时修改或增加复杂性。第二种正则化方法为网络的输入数据添加空间感知,以提高训练稳定性和效率。每个像素的坐标被用作附加特征,打破了神经网络架构的不变性属性。附加特征被证明可以隐式正则化光流估计,强制实现一致的流,同时提高性能和收敛时间。最后,两种正则化方法的结合以互补的方式进一步提高了现有前沿架构的性能,无论是在定量还是定性方面,在流行的光流估计基准数据集上都得到了验证。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/39cb7c28ccf5/sensors-20-03855-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/141df09036e8/sensors-20-03855-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/33d5630c8057/sensors-20-03855-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/46d691320e08/sensors-20-03855-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/ed3ad021b2ef/sensors-20-03855-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/748d43c20d0f/sensors-20-03855-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/19fff353af0c/sensors-20-03855-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/e4b9a87e78af/sensors-20-03855-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/c453d91d8e9b/sensors-20-03855-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/a39502836893/sensors-20-03855-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/d47c96f054aa/sensors-20-03855-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/36ec80c8a079/sensors-20-03855-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/39cb7c28ccf5/sensors-20-03855-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/141df09036e8/sensors-20-03855-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/33d5630c8057/sensors-20-03855-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/46d691320e08/sensors-20-03855-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/ed3ad021b2ef/sensors-20-03855-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/748d43c20d0f/sensors-20-03855-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/19fff353af0c/sensors-20-03855-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/e4b9a87e78af/sensors-20-03855-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/c453d91d8e9b/sensors-20-03855-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/a39502836893/sensors-20-03855-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/d47c96f054aa/sensors-20-03855-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/36ec80c8a079/sensors-20-03855-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fc33/7412516/39cb7c28ccf5/sensors-20-03855-g012.jpg

相似文献

1
Implicit and Explicit Regularization for Optical Flow Estimation.光流估计的隐式和显式正则化。
Sensors (Basel). 2020 Jul 10;20(14):3855. doi: 10.3390/s20143855.
2
Nonlocal sparse and low-rank regularization for optical flow estimation.非局部稀疏和低秩正则化的光流估计。
IEEE Trans Image Process. 2014 Oct;23(10):4527-38. doi: 10.1109/TIP.2014.2352497. Epub 2014 Aug 27.
3
Land Use Classification of the Deep Convolutional Neural Network Method Reducing the Loss of Spatial Features.基于深度卷积神经网络方法的土地利用分类减少空间特征损失
Sensors (Basel). 2019 Jun 21;19(12):2792. doi: 10.3390/s19122792.
4
Lane and Road Marker Semantic Video Segmentation Using Mask Cropping and Optical Flow Estimation.基于掩模裁剪和光流估计的车道和路牌语义视频分割。
Sensors (Basel). 2021 Oct 28;21(21):7156. doi: 10.3390/s21217156.
5
Edge Preserving and Multi-Scale Contextual Neural Network for Salient Object Detection.边缘保持和多尺度上下文神经网络的显著目标检测。
IEEE Trans Image Process. 2018;27(1):121-134. doi: 10.1109/TIP.2017.2756825.
6
Collaborative Deconvolutional Neural Networks for Joint Depth Estimation and Semantic Segmentation.用于联合深度估计和语义分割的协作式反卷积神经网络
IEEE Trans Neural Netw Learn Syst. 2018 Nov;29(11):5655-5666. doi: 10.1109/TNNLS.2017.2787781. Epub 2018 Mar 20.
7
DIO-SLAM: A Dynamic RGB-D SLAM Method Combining Instance Segmentation and Optical Flow.DIO-SLAM:一种结合实例分割与光流的动态RGB-D同步定位与地图构建方法
Sensors (Basel). 2024 Sep 12;24(18):5929. doi: 10.3390/s24185929.
8
Regularization for Unsupervised Learning of Optical Flow.无监督光流学习的正则化。
Sensors (Basel). 2023 Apr 18;23(8):4080. doi: 10.3390/s23084080.
9
DFT-Net: Deep Feature Transformation Based Network for Object Categorization and Part Segmentation in 3-Dimensional Point Clouds.DFT-Net:基于深度特征变换的三维点云目标分类与部件分割网络
Sensors (Basel). 2022 Mar 25;22(7):2512. doi: 10.3390/s22072512.
10
A Fast and Robust Lane Detection Method Based on Semantic Segmentation and Optical Flow Estimation.一种基于语义分割和光流估计的快速鲁棒车道检测方法。
Sensors (Basel). 2021 Jan 8;21(2):400. doi: 10.3390/s21020400.

引用本文的文献

1
Data, Signal and Image Processing and Applications in Sensors.数据、信号和图像处理及其在传感器中的应用。
Sensors (Basel). 2021 May 11;21(10):3323. doi: 10.3390/s21103323.

本文引用的文献

1
Semantically Guided Large Deformation Estimation with Deep Networks.基于语义引导的深度网络大变形估计。
Sensors (Basel). 2020 Mar 4;20(5):1392. doi: 10.3390/s20051392.
2
Large-scale Functional Integration, Rather than Functional Dissociation along Dorsal and Ventral Streams, Underlies Visual Perception and Action.视觉感知和行动的基础是大规模功能整合,而非沿着背侧和腹侧通路的功能分离。
J Cogn Neurosci. 2020 May;32(5):847-861. doi: 10.1162/jocn_a_01527. Epub 2020 Jan 14.
3
Unsupervised Learning for Depth, Ego-Motion, and Optical Flow Estimation Using Coupled Consistency Conditions.
使用耦合一致性条件的深度、自我运动和光流估计的无监督学习
Sensors (Basel). 2019 May 29;19(11):2459. doi: 10.3390/s19112459.
4
Models Matter, So Does Training: An Empirical Study of CNNs for Optical Flow Estimation.模型很重要,训练也很重要:用于光流估计的 CNN 的实证研究。
IEEE Trans Pattern Anal Mach Intell. 2020 Jun;42(6):1408-1423. doi: 10.1109/TPAMI.2019.2894353. Epub 2019 Jan 22.
5
How do the two visual streams interact with each other?这两条视觉信息流是如何相互作用的?
Exp Brain Res. 2017 May;235(5):1297-1308. doi: 10.1007/s00221-017-4917-4. Epub 2017 Mar 2.
6
An investigation of smoothness constraints for the estimation of displacement vector fields from image sequences.基于图像序列的位移矢量场估计的平滑约束研究。
IEEE Trans Pattern Anal Mach Intell. 1986 May;8(5):565-93. doi: 10.1109/tpami.1986.4767833.