• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

学习用于视频超分辨率的深度双注意力网络。

Learning a Deep Dual Attention Network for Video Super-Resolution.

作者信息

Li Feng, Bai Huihui, Zhao Yao

出版信息

IEEE Trans Image Process. 2020 Feb 12. doi: 10.1109/TIP.2020.2972118.

DOI:10.1109/TIP.2020.2972118
PMID:32070957
Abstract

Recently, deep learning based video super-resolution (SR) methods combine the convolutional neural networks (CNN) with motion compensation to estimate a high-resolution (HR) video from its low-resolution (LR) counterpart. However, most previous methods conduct downscaling motion estimation to handle large motions, which can lead to detrimental effects on the accuracy of motion estimation due to the reduction of spatial resolution. Besides, these methods usually treat different types of intermediate features equally, which lack flexibility to emphasize meaningful information for revealing the high-frequency details. In this paper, to solve above issues, we propose a deep dual attention network (DDAN), including a motion compensation network (MCNet) and a SR reconstruction network (ReconNet), to fully exploit the spatio-temporal informative features for accurate video SR. The MCNet progressively learns the optical flow representations to synthesize the motion information across adjacent frames in a pyramid fashion. To decrease the mis-registration errors caused by the optical flow based motion compensation, we extract the detail components of original LR neighboring frames as complementary information for accurate feature extraction. In the ReconNet, we implement dual attention mechanisms on a residual unit and form a residual attention unit to focus on the intermediate informative features for high-frequency details recovery. Extensive experimental results on numerous datasets demonstrate the proposed method can effectively achieve superior performance in terms of quantitative and qualitative assessments compared with state-of-the-art methods.

摘要

最近,基于深度学习的视频超分辨率(SR)方法将卷积神经网络(CNN)与运动补偿相结合,以从低分辨率(LR)视频中估计出高分辨率(HR)视频。然而,大多数先前的方法进行下采样运动估计来处理大运动,由于空间分辨率的降低,这可能会对运动估计的准确性产生不利影响。此外,这些方法通常平等对待不同类型的中间特征,缺乏强调有意义信息以揭示高频细节的灵活性。在本文中,为了解决上述问题,我们提出了一种深度双注意力网络(DDAN),包括一个运动补偿网络(MCNet)和一个SR重建网络(ReconNet),以充分利用时空信息特征来实现准确的视频超分辨率。MCNet逐步学习光流表示,以金字塔方式合成相邻帧之间的运动信息。为了减少基于光流的运动补偿引起的配准误差,我们提取原始LR相邻帧的细节分量作为补充信息,用于准确的特征提取。在ReconNet中,我们在残差单元上实现双注意力机制,并形成一个残差注意力单元,以关注中间信息特征以恢复高频细节。在众多数据集上的大量实验结果表明,与现有方法相比,所提出的方法在定量和定性评估方面都能有效地实现卓越的性能。

相似文献

1
Learning a Deep Dual Attention Network for Video Super-Resolution.学习用于视频超分辨率的深度双注意力网络。
IEEE Trans Image Process. 2020 Feb 12. doi: 10.1109/TIP.2020.2972118.
2
Video Super-Resolution via a Spatio-Temporal Alignment Network.通过时空对齐网络实现视频超分辨率
IEEE Trans Image Process. 2022;31:1761-1773. doi: 10.1109/TIP.2022.3146625. Epub 2022 Feb 8.
3
Multi-Memory Convolutional Neural Network for Video Super-Resolution.用于视频超分辨率的多记忆卷积神经网络。
IEEE Trans Image Process. 2018 Dec 17. doi: 10.1109/TIP.2018.2887017.
4
Deep Video Super-Resolution using HR Optical Flow Estimation.基于高分辨率光流估计的深度视频超分辨率
IEEE Trans Image Process. 2020 Jan 23. doi: 10.1109/TIP.2020.2967596.
5
Learning Temporal Dynamics for Video Super-Resolution: A Deep Learning Approach.学习视频超分辨率的时间动态:一种深度学习方法。
IEEE Trans Image Process. 2018 Mar 30. doi: 10.1109/TIP.2018.2820807.
6
Video Super-Resolution Using Non-Simultaneous Fully Recurrent Convolutional Network.使用非同步全递归卷积网络的视频超分辨率
IEEE Trans Image Process. 2018 Oct 22. doi: 10.1109/TIP.2018.2877334.
7
STDAN: Deformable Attention Network for Space-Time Video Super-Resolution.STDAN:用于时空视频超分辨率的可变形注意力网络。
IEEE Trans Neural Netw Learn Syst. 2024 Aug;35(8):10606-10616. doi: 10.1109/TNNLS.2023.3243029. Epub 2024 Aug 5.
8
DSTAN: A Deformable Spatial-temporal Attention Network with Bidirectional Sequence Feature Refinement for Speckle Noise Removal in Thyroid Ultrasound Video.DSTAN:一种具有双向序列特征细化的可变形时空注意力网络,用于去除甲状腺超声视频中的斑点噪声。
J Imaging Inform Med. 2024 Dec;37(6):3264-3281. doi: 10.1007/s10278-023-00935-5. Epub 2024 Jun 5.
9
MEMC-Net: Motion Estimation and Motion Compensation Driven Neural Network for Video Interpolation and Enhancement.MEMC-Net:用于视频插值与增强的运动估计和运动补偿驱动神经网络。
IEEE Trans Pattern Anal Mach Intell. 2021 Mar;43(3):933-948. doi: 10.1109/TPAMI.2019.2941941. Epub 2021 Feb 4.
10
Video Super-Resolution via Bidirectional Recurrent Convolutional Networks.基于双向循环卷积网络的视频超分辨率重建
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):1015-1028. doi: 10.1109/TPAMI.2017.2701380. Epub 2017 May 4.

引用本文的文献

1
Attention-Based Bi-Prediction Network for Versatile Video Coding (VVC) over 5G Network.基于注意力的双向预测网络在 5G 网络上的通用视频编码 (VVC)中的应用。
Sensors (Basel). 2023 Feb 27;23(5):2631. doi: 10.3390/s23052631.