• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于广谱学习系统的视频时间一致性改进。

Improving Video Temporal Consistency via Broad Learning System.

出版信息

IEEE Trans Cybern. 2022 Jul;52(7):6662-6675. doi: 10.1109/TCYB.2021.3079311. Epub 2022 Jul 4.

DOI:10.1109/TCYB.2021.3079311
PMID:34077381
Abstract

Applying image-based processing methods to original videos on a framewise level breaks the temporal consistency between consecutive frames. Traditional video temporal consistency methods reconstruct an original frame containing flickers from corresponding nonflickering frames, but the inaccurate correspondence realized by optical flow restricts their practical use. In this article, we propose a temporally broad learning system (TBLS), an approach that enforces temporal consistency between frames. We establish the TBLS as a flat network comprising the input data, consisting of an original frame in an original video, a corresponding frame in the temporally inconsistent video on which the image-based technique was applied, and an output frame of the last original frame, as mapped features in feature nodes. Then, we refine extracted features by enhancing the mapped features as enhancement nodes with randomly generated weights. We then connect all extracted features to the output layer with a target weight vector. With the target weight vector, we can minimize the temporal information loss between consecutive frames and the video fidelity loss in the output videos. Finally, we remove the temporal inconsistency in the processed video and output a temporally consistent video. Besides, we propose an alternative incremental learning algorithm based on the increment of the mapped feature nodes, enhancement nodes, or input data to improve learning accuracy by a broad expansion. We demonstrate the superiority of our proposed TBLS by conducting extensive experiments.

摘要

将基于图像的处理方法应用于逐帧的原始视频会打破连续帧之间的时间一致性。传统的视频时间一致性方法从相应的非闪烁帧中重建包含闪烁的原始帧,但光流实现的不准确对应限制了它们的实际应用。在本文中,我们提出了一个时间广泛学习系统(TBLS),这是一种强制帧之间时间一致性的方法。我们将 TBLS 建立为一个平面网络,包括输入数据,由原始视频中的原始帧、应用基于图像的技术的时间不一致视频中的相应帧以及最后原始帧的输出帧组成,这些帧作为特征节点中的映射特征。然后,我们通过增强映射特征作为增强节点并随机生成权重来细化提取的特征。然后,我们将所有提取的特征与输出层连接起来,使用目标权重向量。利用目标权重向量,我们可以最小化连续帧之间的时间信息损失和输出视频中的视频保真度损失。最后,我们消除处理后的视频中的时间不一致性,并输出一个时间一致的视频。此外,我们提出了一种基于映射特征节点、增强节点或输入数据增量的替代增量学习算法,通过广泛扩展来提高学习准确性。我们通过广泛的实验证明了我们提出的 TBLS 的优越性。

相似文献

1
Improving Video Temporal Consistency via Broad Learning System.基于广谱学习系统的视频时间一致性改进。
IEEE Trans Cybern. 2022 Jul;52(7):6662-6675. doi: 10.1109/TCYB.2021.3079311. Epub 2022 Jul 4.
2
Optical flow estimation using temporally oversampled video.使用时间过采样视频的光流估计
IEEE Trans Image Process. 2005 Aug;14(8):1074-87. doi: 10.1109/tip.2005.851688.
3
Full-frame video stabilization with motion inpainting.采用运动修复技术的全帧视频稳定
IEEE Trans Pattern Anal Mach Intell. 2006 Jul;28(7):1150-63. doi: 10.1109/TPAMI.2006.141.
4
Selecting salient frames for spatiotemporal video modeling and segmentation.为时空视频建模与分割选择显著帧。
IEEE Trans Image Process. 2007 Dec;16(12):3035-46. doi: 10.1109/tip.2007.908283.
5
Feature Aggregation With Reinforcement Learning for Video-Based Person Re-Identification.基于强化学习的视频人物再识别特征聚合。
IEEE Trans Neural Netw Learn Syst. 2019 Dec;30(12):3847-3852. doi: 10.1109/TNNLS.2019.2899588. Epub 2019 Mar 12.
6
Semi-supervised video segmentation using tree structured graphical models.基于树状图结构模型的半监督视频分割。
IEEE Trans Pattern Anal Mach Intell. 2013 Nov;35(11):2751-64. doi: 10.1109/TPAMI.2013.54.
7
Scalable gastroscopic video summarization via similar-inhibition dictionary selection.通过相似抑制字典选择实现可扩展的胃镜视频摘要
Artif Intell Med. 2016 Jan;66:1-13. doi: 10.1016/j.artmed.2015.08.006. Epub 2015 Aug 18.
8
Lane and Road Marker Semantic Video Segmentation Using Mask Cropping and Optical Flow Estimation.基于掩模裁剪和光流估计的车道和路牌语义视频分割。
Sensors (Basel). 2021 Oct 28;21(21):7156. doi: 10.3390/s21217156.
9
Network-based H.264/AVC whole frame loss visibility model and frame dropping methods.基于网络的 H.264/AVC 整帧丢失可见性模型和丢帧方法。
IEEE Trans Image Process. 2012 Aug;21(8):3353-63. doi: 10.1109/TIP.2012.2191567. Epub 2012 Mar 21.
10
Video event recognition using kernel methods with multilevel temporal alignment.使用具有多级时间对齐的核方法进行视频事件识别。
IEEE Trans Pattern Anal Mach Intell. 2008 Nov;30(11):1985-97. doi: 10.1109/TPAMI.2008.129.

引用本文的文献

1
Secure Indoor Water Level Monitoring with Temporal Super-Resolution and Enhanced Yolov5.基于时间超分辨率和增强型Yolov5的安全室内水位监测
Sensors (Basel). 2025 Apr 30;25(9):2835. doi: 10.3390/s25092835.
2
A Hybrid Approach for Image Acquisition Methods Based on Feature-Based Image Registration.一种基于特征的图像配准的图像采集方法的混合方法。
J Imaging. 2024 Sep 14;10(9):228. doi: 10.3390/jimaging10090228.
3
Motion Capture Technology in Sports Scenarios: A Survey.运动场景中的运动捕捉技术:综述。
Sensors (Basel). 2024 May 6;24(9):2947. doi: 10.3390/s24092947.
4
Object-Oriented and Visual-Based Localization in Urban Environments.城市环境中基于对象和视觉的定位
Sensors (Basel). 2024 Mar 21;24(6):2014. doi: 10.3390/s24062014.