• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

DSNet:一种用于视频摘要的灵活检测到摘要网络。

DSNet: A Flexible Detect-to-Summarize Network for Video Summarization.

作者信息

Zhu Wencheng, Lu Jiwen, Li Jiahao, Zhou Jie

出版信息

IEEE Trans Image Process. 2021;30:948-962. doi: 10.1109/TIP.2020.3039886. Epub 2020 Dec 8.

DOI:10.1109/TIP.2020.3039886
PMID:33259299
Abstract

In this paper, we propose a Detect-to-Summarize network (DSNet) framework for supervised video summarization. Our DSNet contains anchor-based and anchor-free counterparts. The anchor-based method generates temporal interest proposals to determine and localize the representative contents of video sequences, while the anchor-free method eliminates the pre-defined temporal proposals and directly predicts the importance scores and segment locations. Different from existing supervised video summarization methods which formulate video summarization as a regression problem without temporal consistency and integrity constraints, our interest detection framework is the first attempt to leverage temporal consistency via the temporal interest detection formulation. Specifically, in the anchor-based approach, we first provide a dense sampling of temporal interest proposals with multi-scale intervals that accommodate interest variations in length, and then extract their long-range temporal features for interest proposal location regression and importance prediction. Notably, positive and negative segments are both assigned for the correctness and completeness information of the generated summaries. In the anchor-free approach, we alleviate drawbacks of temporal proposals by directly predicting importance scores of video frames and segment locations. Particularly, the interest detection framework can be flexibly plugged into off-the-shelf supervised video summarization methods. We evaluate the anchor-based and anchor-free approaches on the SumMe and TVSum datasets. Experimental results clearly validate the effectiveness of the anchor-based and anchor-free approaches.

摘要

在本文中,我们提出了一种用于有监督视频摘要的检测到摘要网络(DSNet)框架。我们的DSNet包含基于锚点和无锚点的对应方法。基于锚点的方法生成时间兴趣提议,以确定和定位视频序列的代表性内容,而无锚点的方法则消除了预定义的时间提议,并直接预测重要性得分和片段位置。与现有的将视频摘要表述为无时间一致性和完整性约束的回归问题的有监督视频摘要方法不同,我们的兴趣检测框架是首次尝试通过时间兴趣检测公式利用时间一致性。具体而言,在基于锚点的方法中,我们首先提供具有多尺度间隔的时间兴趣提议的密集采样,以适应长度上的兴趣变化,然后提取它们的长程时间特征,用于兴趣提议位置回归和重要性预测。值得注意的是,为生成的摘要的正确性和完整性信息同时分配了正片段和负片段。在无锚点的方法中,我们通过直接预测视频帧的重要性得分和片段位置来减轻时间提议的缺点。特别是,兴趣检测框架可以灵活地插入现成的有监督视频摘要方法中。我们在SumMe和TVSum数据集上评估了基于锚点和无锚点的方法。实验结果清楚地验证了基于锚点和无锚点方法的有效性。

相似文献

1
DSNet: A Flexible Detect-to-Summarize Network for Video Summarization.DSNet:一种用于视频摘要的灵活检测到摘要网络。
IEEE Trans Image Process. 2021;30:948-962. doi: 10.1109/TIP.2020.3039886. Epub 2020 Dec 8.
2
Interp-SUM: Unsupervised Video Summarization with Piecewise Linear Interpolation.Interp-SUM:基于分段线性插值的无监督视频摘要。
Sensors (Basel). 2021 Jul 2;21(13):4562. doi: 10.3390/s21134562.
3
Relational Reasoning Over Spatial-Temporal Graphs for Video Summarization.用于视频摘要的时空图关系推理
IEEE Trans Image Process. 2022;31:3017-3031. doi: 10.1109/TIP.2022.3163855. Epub 2022 Apr 11.
4
Unsupervised Video Summarization Based on Deep Reinforcement Learning with Interpolation.基于深度强化学习与插值的无监督视频摘要。
Sensors (Basel). 2023 Mar 23;23(7):3384. doi: 10.3390/s23073384.
5
Video Summarization With Spatiotemporal Vision Transformer.基于时空视觉Transformer 的视频摘要
IEEE Trans Image Process. 2023;32:3013-3026. doi: 10.1109/TIP.2023.3275069. Epub 2023 May 26.
6
A Video Summarization Model Based on Deep Reinforcement Learning with Long-Term Dependency.基于深度强化学习和长期依赖的视频摘要模型。
Sensors (Basel). 2022 Oct 10;22(19):7689. doi: 10.3390/s22197689.
7
AudioVisual Video Summarization.视听视频摘要
IEEE Trans Neural Netw Learn Syst. 2023 Aug;34(8):5181-5188. doi: 10.1109/TNNLS.2021.3119969. Epub 2023 Aug 4.
8
Property-Constrained Dual Learning for Video Summarization.用于视频摘要的属性约束对偶学习
IEEE Trans Neural Netw Learn Syst. 2020 Oct;31(10):3989-4000. doi: 10.1109/TNNLS.2019.2951680. Epub 2019 Dec 5.
9
Unleashing the Power of Contrastive Learning for Zero-Shot Video Summarization.释放对比学习在零样本视频摘要中的力量。
J Imaging. 2024 Sep 14;10(9):229. doi: 10.3390/jimaging10090229.
10
User-Ranking Video Summarization with Multi-Stage Spatio-Temporal Representation.基于多阶段时空表征的用户排序视频摘要
IEEE Trans Image Process. 2018 Dec 21. doi: 10.1109/TIP.2018.2889265.

引用本文的文献

1
Key frame extraction algorithm for surveillance videos using an evolutionary approach.基于进化方法的监控视频关键帧提取算法
Sci Rep. 2025 Jan 2;15(1):536. doi: 10.1038/s41598-024-84324-0.
2
Behavioral profiling for adaptive video summarization: From generalization to personalization.用于自适应视频摘要的行为分析:从通用化到个性化
MethodsX. 2024 Jun 14;13:102780. doi: 10.1016/j.mex.2024.102780. eCollection 2024 Dec.