• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一种适用于各种监控环境的统一递归视频对象分割框架。

An Unified Recurrent Video Object Segmentation Framework for Various Surveillance Environments.

作者信息

Patil Prashant W, Dudhane Akshay, Kulkarni Ashutosh, Murala Subrahmanyam, Gonde Anil Balaji, Gupta Sunil

出版信息

IEEE Trans Image Process. 2021;30:7889-7902. doi: 10.1109/TIP.2021.3108405. Epub 2021 Sep 20.

DOI:10.1109/TIP.2021.3108405
PMID:34478367
Abstract

Moving object segmentation (MOS) in videos received considerable attention because of its broad security-based applications like robotics, outdoor video surveillance, self-driving cars, etc. The current prevailing algorithms highly depend on additional trained modules for other applications or complicated training procedures or neglect the inter-frame spatio-temporal structural dependencies. To address these issues, a simple, robust, and effective unified recurrent edge aggregation approach is proposed for MOS, in which additional trained modules or fine-tuning on a test video frame(s) are not required. Here, a recurrent edge aggregation module (REAM) is proposed to extract effective foreground relevant features capturing spatio-temporal structural dependencies with encoder and respective decoder features connected recurrently from previous frame. These REAM features are then connected to a decoder through skip connections for comprehensive learning named as temporal information propagation. Further, the motion refinement block with multi-scale dense residual is proposed to combine the features from the optical flow encoder stream and the last REAM module for holistic feature learning. Finally, these holistic features and REAM features are given to the decoder block for segmentation. To guide the decoder block, previous frame output with respective scales is utilized. The different configurations of training-testing techniques are examined to evaluate the performance of the proposed method. Specifically, outdoor videos often suffer from constrained visibility due to different environmental conditions and other small particles in the air that scatter the light in the atmosphere. Thus, comprehensive result analysis is conducted on six benchmark video datasets with different surveillance environments. We demonstrate that the proposed method outperforms the state-of-the-art methods for MOS without any pre-trained module, fine-tuning on the test video frame(s) or complicated training.

摘要

视频中的运动目标分割(MOS)因其在机器人技术、户外视频监控、自动驾驶汽车等基于安全的广泛应用而受到广泛关注。当前流行的算法高度依赖于用于其他应用的额外训练模块或复杂的训练过程,或者忽略了帧间时空结构依赖性。为了解决这些问题,本文提出了一种简单、稳健且有效的统一递归边缘聚合方法用于MOS,该方法无需额外的训练模块或在测试视频帧上进行微调。在此,提出了一种递归边缘聚合模块(REAM),通过将编码器和各自的解码器特征与前一帧递归连接来提取有效的前景相关特征,从而捕捉时空结构依赖性。然后,这些REAM特征通过跳跃连接连接到解码器,进行全面学习,称为时间信息传播。此外,还提出了具有多尺度密集残差的运动细化模块,将光流编码器流和最后一个REAM模块的特征相结合,进行整体特征学习。最后,将这些整体特征和REAM特征输入到解码器模块进行分割。为了指导解码器模块,利用了具有相应尺度的前一帧输出。研究了不同的训练 - 测试技术配置,以评估所提出方法的性能。具体而言,户外视频由于不同的环境条件和空气中散射光线的其他小颗粒,常常存在能见度受限的问题。因此,在六个具有不同监控环境的基准视频数据集上进行了全面的结果分析。我们证明,所提出的方法在没有任何预训练模块、不在测试视频帧上进行微调或采用复杂训练的情况下,优于当前最先进的MOS方法。

相似文献

1
An Unified Recurrent Video Object Segmentation Framework for Various Surveillance Environments.一种适用于各种监控环境的统一递归视频对象分割框架。
IEEE Trans Image Process. 2021;30:7889-7902. doi: 10.1109/TIP.2021.3108405. Epub 2021 Sep 20.
2
A spatio-temporal network for video semantic segmentation in surgical videos.用于手术视频中视频语义分割的时空网络。
Int J Comput Assist Radiol Surg. 2024 Feb;19(2):375-382. doi: 10.1007/s11548-023-02971-6. Epub 2023 Jun 22.
3
Hierarchical Co-Attention Propagation Network for Zero-Shot Video Object Segmentation.层次化协同注意传播网络的零样本视频对象分割。
IEEE Trans Image Process. 2023;32:2348-2359. doi: 10.1109/TIP.2023.3267244. Epub 2023 Apr 25.
4
Language-Aware Spatial-Temporal Collaboration for Referring Video Segmentation.面向视频分割的语言感知时空协作
IEEE Trans Pattern Anal Mach Intell. 2023 Jul;45(7):8646-8659. doi: 10.1109/TPAMI.2023.3235720. Epub 2023 Jun 5.
5
DSTAN: A Deformable Spatial-temporal Attention Network with Bidirectional Sequence Feature Refinement for Speckle Noise Removal in Thyroid Ultrasound Video.DSTAN:一种具有双向序列特征细化的可变形时空注意力网络,用于去除甲状腺超声视频中的斑点噪声。
J Imaging Inform Med. 2024 Dec;37(6):3264-3281. doi: 10.1007/s10278-023-00935-5. Epub 2024 Jun 5.
6
Motion-Guided Cascaded Refinement Network for Video Object Segmentation.用于视频对象分割的运动引导级联细化网络
IEEE Trans Pattern Anal Mach Intell. 2020 Aug;42(8):1957-1967. doi: 10.1109/TPAMI.2019.2906175. Epub 2019 Mar 19.
7
Feature Aggregation and Propagation Network for Camouflaged Object Detection.用于伪装目标检测的特征聚合与传播网络
IEEE Trans Image Process. 2022;31:7036-7047. doi: 10.1109/TIP.2022.3217695. Epub 2022 Nov 14.
8
Bilateral attention decoder: A lightweight decoder for real-time semantic segmentation.双边注意解码器:用于实时语义分割的轻量级解码器。
Neural Netw. 2021 May;137:188-199. doi: 10.1016/j.neunet.2021.01.021. Epub 2021 Jan 30.
9
Referring Segmentation in Images and Videos With Cross-Modal Self-Attention Network.基于跨模态自注意力网络的图像和视频指代分割
IEEE Trans Pattern Anal Mach Intell. 2022 Jul;44(7):3719-3732. doi: 10.1109/TPAMI.2021.3054384. Epub 2022 Jun 3.
10
Spatio-Temporal Attention Model for Foreground Detection in Cross-Scene Surveillance Videos.跨场景监控视频中前景检测的时空注意模型。
Sensors (Basel). 2019 Nov 24;19(23):5142. doi: 10.3390/s19235142.

引用本文的文献

1
Intelligent Object Tracking with an Automatic Image Zoom Algorithm for a Camera Sensing Surveillance System.基于自动图像缩放算法的智能目标跟踪摄像头感测监控系统。
Sensors (Basel). 2022 Nov 14;22(22):8791. doi: 10.3390/s22228791.