• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于背景模型补全的运动相机显著性检测

Saliency Detection with Moving Camera via Background Model Completion.

作者信息

Zhang Yu-Pei, Chan Kwok-Leung

机构信息

Department of Electrical Engineering, City University of Hong Kong, Hong Kong, China.

出版信息

Sensors (Basel). 2021 Dec 15;21(24):8374. doi: 10.3390/s21248374.

DOI:10.3390/s21248374
PMID:34960461
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8707474/
Abstract

Detecting saliency in videos is a fundamental step in many computer vision systems. Saliency is the significant target(s) in the video. The object of interest is further analyzed for high-level applications. The segregation of saliency and the background can be made if they exhibit different visual cues. Therefore, saliency detection is often formulated as background subtraction. However, saliency detection is challenging. For instance, dynamic background can result in false positive errors. In another scenario, camouflage will result in false negative errors. With moving cameras, the captured scenes are even more complicated to handle. We propose a new framework, called saliency detection via background model completion (SD-BMC), that comprises a background modeler and a deep learning background/foreground segmentation network. The background modeler generates an initial clean background image from a short image sequence. Based on the idea of video completion, a good background frame can be synthesized with the co-existence of changing background and moving objects. We adopt the background/foreground segmenter, which was pre-trained with a specific video dataset. It can also detect saliency in unseen videos. The background modeler can adjust the background image dynamically when the background/foreground segmenter output deteriorates during processing a long video. To the best of our knowledge, our framework is the first one to adopt video completion for background modeling and saliency detection in videos captured by moving cameras. The F-measure results, obtained from the pan-tilt-zoom (PTZ) videos, show that our proposed framework outperforms some deep learning-based background subtraction models by 11% or more. With more challenging videos, our framework also outperforms many high-ranking background subtraction methods by more than 3%.

摘要

检测视频中的显著区域是许多计算机视觉系统的基本步骤。显著区域是视频中的重要目标。为了进行高级应用,需要进一步分析感兴趣的对象。如果显著区域和背景呈现出不同的视觉线索,就可以将它们区分开来。因此,显著区域检测通常被表述为背景减法。然而,显著区域检测具有挑战性。例如,动态背景可能导致误报错误。在另一种情况下,伪装会导致漏报错误。对于移动摄像头,捕获的场景更难处理。我们提出了一种新的框架,称为通过背景模型完成的显著区域检测(SD-BMC),它包括一个背景建模器和一个深度学习背景/前景分割网络。背景建模器从一个短图像序列生成初始的干净背景图像。基于视频完成的思想,可以在变化的背景和移动物体共存的情况下合成一个好的背景帧。我们采用经过特定视频数据集预训练的背景/前景分割器。它也可以在未见过的视频中检测显著区域。当背景/前景分割器在处理长视频时输出质量下降时,背景建模器可以动态调整背景图像。据我们所知,我们的框架是第一个在移动摄像头捕获的视频中采用视频完成进行背景建模和显著区域检测的框架。从云台变焦(PTZ)视频获得的F值结果表明,我们提出的框架比一些基于深度学习的背景减法模型性能高出11%或更多。对于更具挑战性的视频,我们的框架也比许多高级背景减法方法性能高出3%以上。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/76be/8707474/0dbdc9114d79/sensors-21-08374-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/76be/8707474/4d7cfa44868e/sensors-21-08374-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/76be/8707474/565e373d88fd/sensors-21-08374-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/76be/8707474/2f95d021782d/sensors-21-08374-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/76be/8707474/e624c62cd601/sensors-21-08374-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/76be/8707474/ce7120cf5f54/sensors-21-08374-g005a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/76be/8707474/e09055f443a2/sensors-21-08374-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/76be/8707474/0dbdc9114d79/sensors-21-08374-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/76be/8707474/4d7cfa44868e/sensors-21-08374-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/76be/8707474/565e373d88fd/sensors-21-08374-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/76be/8707474/2f95d021782d/sensors-21-08374-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/76be/8707474/e624c62cd601/sensors-21-08374-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/76be/8707474/ce7120cf5f54/sensors-21-08374-g005a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/76be/8707474/e09055f443a2/sensors-21-08374-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/76be/8707474/0dbdc9114d79/sensors-21-08374-g007.jpg

相似文献

1
Saliency Detection with Moving Camera via Background Model Completion.基于背景模型补全的运动相机显著性检测
Sensors (Basel). 2021 Dec 15;21(24):8374. doi: 10.3390/s21248374.
2
Deep Features Homography Transformation Fusion Network-A Universal Foreground Segmentation Algorithm for PTZ Cameras and a Comparative Study.深度特征单应性变换融合网络——一种用于云台摄像机的通用前景分割算法及比较研究
Sensors (Basel). 2020 Jun 17;20(12):3420. doi: 10.3390/s20123420.
3
Video Salient Object Detection via Fully Convolutional Networks.基于全卷积网络的视频显著目标检测
IEEE Trans Image Process. 2018;27(1):38-49. doi: 10.1109/TIP.2017.2754941.
4
Hierarchical ensemble of background models for PTZ-based video surveillance.基于 PTZ 的视频监控的分层背景模型集成。
IEEE Trans Cybern. 2015 Jan;45(1):89-102. doi: 10.1109/TCYB.2014.2320493. Epub 2014 May 20.
5
SCOM: Spatiotemporal Constrained Optimization for Salient Object Detection.SCOM:显著目标检测的时空约束优化。
IEEE Trans Image Process. 2018 Jul;27(7):3345-3357. doi: 10.1109/TIP.2018.2813165.
6
Segmentation in Weakly Labeled Videos via a Semantic Ranking and Optical Warping Network.通过语义排序和光流变形网络对弱标注视频进行分割
IEEE Trans Image Process. 2018 May 16. doi: 10.1109/TIP.2018.2834221.
7
A Benchmark Dataset and Saliency-Guided Stacked Autoencoders for Video-Based Salient Object Detection.基于视频的显著目标检测的基准数据集和显著引导堆叠自动编码器。
IEEE Trans Image Process. 2018 Jan;27(1):349-364. doi: 10.1109/TIP.2017.2762594. Epub 2017 Oct 12.
8
Fusing Self-Organized Neural Network and Keypoint Clustering for Localized Real-Time Background Subtraction.基于自组织神经网络和关键点聚类的本地化实时背景减除。
Int J Neural Syst. 2020 Apr;30(4):2050016. doi: 10.1142/S0129065720500161. Epub 2020 Mar 2.
9
A parallel spatiotemporal saliency and discriminative online learning method for visual target tracking in aerial videos.一种用于航空视频视觉目标跟踪的并行时空显著性与判别式在线学习方法。
PLoS One. 2018 Feb 13;13(2):e0192246. doi: 10.1371/journal.pone.0192246. eCollection 2018.
10
A Multimodal Saliency Model for Videos with High Audio-Visual Correspondence.一种用于具有高视听对应性视频的多模态显著性模型。
IEEE Trans Image Process. 2020 Jan 17. doi: 10.1109/TIP.2020.2966082.

引用本文的文献

1
Moving Object Detection in Freely Moving Camera via Global Motion Compensation and Local Spatial Information Fusion.基于全局运动补偿和局部空间信息融合的自由移动相机中的运动目标检测
Sensors (Basel). 2024 Apr 30;24(9):2859. doi: 10.3390/s24092859.
2
Video Sequence Segmentation Based on K-Means in Air-Gap Data Transmission for a Cluttered Environment.基于 K-Means 的杂乱环境中空隙数据传输视频序列分割。
Sensors (Basel). 2023 Jan 6;23(2):665. doi: 10.3390/s23020665.

本文引用的文献

1
Deep Features Homography Transformation Fusion Network-A Universal Foreground Segmentation Algorithm for PTZ Cameras and a Comparative Study.深度特征单应性变换融合网络——一种用于云台摄像机的通用前景分割算法及比较研究
Sensors (Basel). 2020 Jun 17;20(12):3420. doi: 10.3390/s20123420.
2
Segmentation of Moving Objects by Long Term Video Analysis.基于长期视频分析的运动目标分割。
IEEE Trans Pattern Anal Mach Intell. 2014 Jun;36(6):1187-200. doi: 10.1109/TPAMI.2013.242.
3
SuBSENSE: a universal change detection method with local adaptive sensitivity.
SuBSENSE:一种具有局部自适应灵敏度的通用变化检测方法。
IEEE Trans Image Process. 2015 Jan;24(1):359-73. doi: 10.1109/TIP.2014.2378053. Epub 2014 Dec 4.
4
A novel video dataset for change detection benchmarking.用于变化检测基准测试的新型视频数据集。
IEEE Trans Image Process. 2014 Nov;23(11):4663-79. doi: 10.1109/TIP.2014.2346013. Epub 2014 Aug 7.
5
A self-organizing approach to background subtraction for visual surveillance applications.一种用于视觉监控应用的背景减除自组织方法。
IEEE Trans Image Process. 2008 Jul;17(7):1168-77. doi: 10.1109/TIP.2008.924285.