• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

EHNet:用于图像去模糊的具有双重注意力的高效混合网络。

EHNet: Efficient Hybrid Network with Dual Attention for Image Deblurring.

作者信息

Ho Quoc-Thien, Duong Minh-Thien, Lee Seongsoo, Hong Min-Cheol

机构信息

Department of Information and Telecommunication Engineering, Soongsil University, Seoul 06978, Republic of Korea.

Department of Automatic Control, Ho Chi Minh City University of Technology and Education, Ho Chi Minh City 70000, Vietnam.

出版信息

Sensors (Basel). 2024 Oct 10;24(20):6545. doi: 10.3390/s24206545.

DOI:10.3390/s24206545
PMID:39460026
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11511264/
Abstract

The motion of an object or camera platform makes the acquired image blurred. This degradation is a major reason to obtain a poor-quality image from an imaging sensor. Therefore, developing an efficient deep-learning-based image processing method to remove the blur artifact is desirable. Deep learning has recently demonstrated significant efficacy in image deblurring, primarily through convolutional neural networks (CNNs) and Transformers. However, the limited receptive fields of CNNs restrict their ability to capture long-range structural dependencies. In contrast, Transformers excel at modeling these dependencies, but they are computationally expensive for high-resolution inputs and lack the appropriate inductive bias. To overcome these challenges, we propose an Efficient Hybrid Network (EHNet) that employs CNN encoders for local feature extraction and Transformer decoders with a dual-attention module to capture spatial and channel-wise dependencies. This synergy facilitates the acquisition of rich contextual information for high-quality image deblurring. Additionally, we introduce the Simple Feature-Embedding Module (SFEM) to replace the pointwise and depthwise convolutions to generate simplified embedding features in the self-attention mechanism. This innovation substantially reduces computational complexity and memory usage while maintaining overall performance. Finally, through comprehensive experiments, our compact model yields promising quantitative and qualitative results for image deblurring on various benchmark datasets.

摘要

物体或相机平台的运动会使采集到的图像模糊。这种退化是从成像传感器获得低质量图像的主要原因。因此,开发一种基于深度学习的高效图像处理方法来去除模糊伪像是很有必要的。深度学习最近在图像去模糊方面显示出显著的效果,主要是通过卷积神经网络(CNN)和Transformer。然而,CNN有限的感受野限制了它们捕捉长距离结构依赖性的能力。相比之下,Transformer擅长对这些依赖性进行建模,但对于高分辨率输入来说计算成本很高,并且缺乏适当的归纳偏差。为了克服这些挑战,我们提出了一种高效混合网络(EHNet),它采用CNN编码器进行局部特征提取,并使用带有双注意力模块的Transformer解码器来捕捉空间和通道方向的依赖性。这种协同作用有助于获取丰富的上下文信息以进行高质量的图像去模糊。此外,我们引入了简单特征嵌入模块(SFEM)来取代逐点卷积和深度卷积,以便在自注意力机制中生成简化的嵌入特征。这一创新在保持整体性能的同时,大幅降低了计算复杂度和内存使用。最后,通过全面的实验,我们的紧凑模型在各种基准数据集上的图像去模糊任务中产生了有希望的定量和定性结果。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/da0f0d083cb7/sensors-24-06545-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/0edac7813f77/sensors-24-06545-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/c5fbdb10075b/sensors-24-06545-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/cccbc4febd7c/sensors-24-06545-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/d53f6336a32a/sensors-24-06545-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/eaa6233f1d22/sensors-24-06545-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/d79368c466ad/sensors-24-06545-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/efd33954ad4d/sensors-24-06545-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/d6bb6c9287ec/sensors-24-06545-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/653cbc29fa50/sensors-24-06545-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/76036c0505c1/sensors-24-06545-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/da0f0d083cb7/sensors-24-06545-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/0edac7813f77/sensors-24-06545-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/c5fbdb10075b/sensors-24-06545-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/cccbc4febd7c/sensors-24-06545-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/d53f6336a32a/sensors-24-06545-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/eaa6233f1d22/sensors-24-06545-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/d79368c466ad/sensors-24-06545-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/efd33954ad4d/sensors-24-06545-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/d6bb6c9287ec/sensors-24-06545-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/653cbc29fa50/sensors-24-06545-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/76036c0505c1/sensors-24-06545-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d08/11511264/da0f0d083cb7/sensors-24-06545-g011.jpg

相似文献

1
EHNet: Efficient Hybrid Network with Dual Attention for Image Deblurring.EHNet:用于图像去模糊的具有双重注意力的高效混合网络。
Sensors (Basel). 2024 Oct 10;24(20):6545. doi: 10.3390/s24206545.
2
An Efficient Image Deblurring Network with a Hybrid Architecture.一种具有混合架构的高效图像去模糊网络。
Sensors (Basel). 2023 Aug 18;23(16):7260. doi: 10.3390/s23167260.
3
HTC-retina: A hybrid retinal diseases classification model using transformer-Convolutional Neural Network from optical coherence tomography images.HTC-retina:一种使用来自光学相干断层扫描图像的变压器-卷积神经网络的混合视网膜疾病分类模型。
Comput Biol Med. 2024 Aug;178:108726. doi: 10.1016/j.compbiomed.2024.108726. Epub 2024 Jun 9.
4
SharpFormer: Learning Local Feature Preserving Global Representations for Image Deblurring.SharpFormer:用于图像去模糊的学习局部特征保持全局表示。
IEEE Trans Image Process. 2023;32:2857-2866. doi: 10.1109/TIP.2023.3251029. Epub 2023 May 26.
5
VSmTrans: A hybrid paradigm integrating self-attention and convolution for 3D medical image segmentation.VSmTrans:一种融合自注意力机制和卷积的 3D 医学图像分割混合范式。
Med Image Anal. 2024 Dec;98:103295. doi: 10.1016/j.media.2024.103295. Epub 2024 Aug 24.
6
Improved UNet with Attention for Medical Image Segmentation.基于注意力机制的改进型 UNet 用于医学图像分割。
Sensors (Basel). 2023 Oct 20;23(20):8589. doi: 10.3390/s23208589.
7
MMNet: A Mixing Module Network for Polyp Segmentation.MMNet:一种用于息肉分割的混合模块网络。
Sensors (Basel). 2023 Aug 18;23(16):7258. doi: 10.3390/s23167258.
8
TGDAUNet: Transformer and GCNN based dual-branch attention UNet for medical image segmentation.TGDAUNet:基于 Transformer 和 GCNN 的双分支注意力 U-Net 用于医学图像分割。
Comput Biol Med. 2023 Dec;167:107583. doi: 10.1016/j.compbiomed.2023.107583. Epub 2023 Oct 21.
9
CVTrack: Combined Convolutional Neural Network and Vision Transformer Fusion Model for Visual Tracking.CVTrack:用于视觉跟踪的卷积神经网络与视觉Transformer融合模型
Sensors (Basel). 2024 Jan 3;24(1):274. doi: 10.3390/s24010274.
10
A dual-branch and dual attention transformer and CNN hybrid network for ultrasound image segmentation.一种用于超声图像分割的双分支双注意力Transformer与CNN混合网络。
Front Physiol. 2024 Sep 27;15:1432987. doi: 10.3389/fphys.2024.1432987. eCollection 2024.

引用本文的文献

1
Adaptive Image Deblurring Convolutional Neural Network with Meta-Tuning.具有元调优的自适应图像去模糊卷积神经网络
Sensors (Basel). 2025 Aug 21;25(16):5211. doi: 10.3390/s25165211.

本文引用的文献

1
Multi-Branch Network for Color Image Denoising Using Dilated Convolution and Attention Mechanisms.基于空洞卷积和注意力机制的彩色图像去噪多分支网络
Sensors (Basel). 2024 Jun 3;24(11):3608. doi: 10.3390/s24113608.
2
CNN-Based Facial Expression Recognition with Simultaneous Consideration of Inter-Class and Intra-Class Variations.基于卷积神经网络的同时考虑类间和类内变化的面部表情识别
Sensors (Basel). 2023 Dec 6;23(24):9658. doi: 10.3390/s23249658.
3
Image Restoration via Frequency Selection.通过频率选择进行图像恢复
IEEE Trans Pattern Anal Mach Intell. 2024 Feb;46(2):1093-1108. doi: 10.1109/TPAMI.2023.3330416. Epub 2024 Jan 8.
4
An Efficient Image Deblurring Network with a Hybrid Architecture.一种具有混合架构的高效图像去模糊网络。
Sensors (Basel). 2023 Aug 18;23(16):7260. doi: 10.3390/s23167260.
5
SharpFormer: Learning Local Feature Preserving Global Representations for Image Deblurring.SharpFormer:用于图像去模糊的学习局部特征保持全局表示。
IEEE Trans Image Process. 2023;32:2857-2866. doi: 10.1109/TIP.2023.3251029. Epub 2023 May 26.
6
BANet: A Blur-Aware Attention Network for Dynamic Scene Deblurring.BANet:一种用于动态场景去模糊的模糊感知注意力网络。
IEEE Trans Image Process. 2022;31:6789-6799. doi: 10.1109/TIP.2022.3216216. Epub 2022 Oct 28.
7
Low-Light Image Enhancement Using Hybrid Deep-Learning and Mixed-Norm Loss Functions.基于混合深度学习和混合范数损失函数的低光照图像增强
Sensors (Basel). 2022 Sep 13;22(18):6904. doi: 10.3390/s22186904.
8
Deblurring Images via Dark Channel Prior.通过暗通道先验去模糊图像
IEEE Trans Pattern Anal Mach Intell. 2018 Oct;40(10):2315-2328. doi: 10.1109/TPAMI.2017.2753804. Epub 2017 Sep 22.
9
Image quality assessment: from error visibility to structural similarity.图像质量评估:从误差可见性到结构相似性。
IEEE Trans Image Process. 2004 Apr;13(4):600-12. doi: 10.1109/tip.2003.819861.