• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于遥感图像超分辨率的不确定性驱动混合卷积与变压器网络

Uncertainty-driven mixture convolution and transformer network for remote sensing image super-resolution.

作者信息

Zhang Xiaomin

机构信息

College of Internet of Things and Artificial Intelligence, Fujian Polytechnic of Information Technology, Fuzhou, 350003, Fujian, China.

出版信息

Sci Rep. 2024 Apr 24;14(1):9435. doi: 10.1038/s41598-024-59384-x.

DOI:10.1038/s41598-024-59384-x
PMID:38658596
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11043356/
Abstract

Recently, convolutional neural networks (CNNs) and Transformer-based Networks have exhibited remarkable prowess in the realm of remote sensing image super-resolution (RSISR), delivering promising results in the field. Nevertheless, the effective fusion of the inductive bias inherent in CNNs and the long-range modeling capabilities encapsulated within the Transformer architecture remains a relatively uncharted terrain in the context of RSISR endeavors. Accordingly, we propose an uncertainty-driven mixture convolution and transformer network (UMCTN) to earn a performance promotion. Specifically, to acquire multi-scale and hierarchical features, UMCTN adopts a U-shape architecture. Utilizing the dual-view aggregation block (DAB) based residual dual-view aggregation group (RDAG) in both encoder and decoder, we solely introduce a pioneering dense-sparse transformer group (DSTG) into the latent layer. This design effectively eradicates the considerable quadratic complexity inherent in vanilla Transformer structures. Moreover, we introduce a novel uncertainty-driven Loss (UDL) to steer the network's attention towards pixels exhibiting significant variance. The primary objective is to elevate the reconstruction quality specifically in texture and edge regions. Experimental outcomes on the UCMerced LandUse and AID datasets unequivocally affirm that UMCTN achieves state-of-the-art performance in comparison to presently prevailing methodologies.

摘要

最近,卷积神经网络(CNN)和基于Transformer的网络在遥感图像超分辨率(RSISR)领域展现出了卓越的能力,在该领域取得了令人瞩目的成果。然而,在RSISR的背景下,将CNN固有的归纳偏差与Transformer架构所具备的远程建模能力进行有效融合,仍然是一个相对未知的领域。因此,我们提出了一种不确定性驱动的混合卷积与Transformer网络(UMCTN)来提升性能。具体而言,为了获取多尺度和分层特征,UMCTN采用了U形架构。通过在编码器和解码器中使用基于双视图聚合块(DAB)的残差双视图聚合组(RDAG),我们仅在潜在层引入了一个开创性的密集-稀疏Transformer组(DSTG)。这种设计有效地消除了普通Transformer结构中固有的相当大的二次复杂性。此外,我们引入了一种新颖的不确定性驱动损失(UDL),以引导网络关注具有显著方差的像素。主要目标是特别提高纹理和边缘区域的重建质量。在UCMerced LandUse和AID数据集上的实验结果明确证实,与目前流行的方法相比,UMCTN实现了领先的性能。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/49d9528ea74c/41598_2024_59384_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/e9991fcc73a7/41598_2024_59384_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/e79e9ac747aa/41598_2024_59384_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/52fed5f0849e/41598_2024_59384_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/51316c0d7500/41598_2024_59384_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/92fcd5e8f584/41598_2024_59384_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/7f88e2b3d6ae/41598_2024_59384_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/2b38e14fdfa5/41598_2024_59384_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/a9e8aaabfafe/41598_2024_59384_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/6e6b42699447/41598_2024_59384_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/49d9528ea74c/41598_2024_59384_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/e9991fcc73a7/41598_2024_59384_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/e79e9ac747aa/41598_2024_59384_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/52fed5f0849e/41598_2024_59384_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/51316c0d7500/41598_2024_59384_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/92fcd5e8f584/41598_2024_59384_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/7f88e2b3d6ae/41598_2024_59384_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/2b38e14fdfa5/41598_2024_59384_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/a9e8aaabfafe/41598_2024_59384_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/6e6b42699447/41598_2024_59384_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a329/11043356/49d9528ea74c/41598_2024_59384_Fig10_HTML.jpg

相似文献

1
Uncertainty-driven mixture convolution and transformer network for remote sensing image super-resolution.用于遥感图像超分辨率的不确定性驱动混合卷积与变压器网络
Sci Rep. 2024 Apr 24;14(1):9435. doi: 10.1038/s41598-024-59384-x.
2
Context-aware lightweight remote-sensing image super-resolution network.上下文感知轻量级遥感图像超分辨率网络。
Front Neurorobot. 2023 Jun 23;17:1220166. doi: 10.3389/fnbot.2023.1220166. eCollection 2023.
3
Multi-attention fusion transformer for single-image super-resolution.用于单图像超分辨率的多注意力融合变压器
Sci Rep. 2024 May 3;14(1):10222. doi: 10.1038/s41598-024-60579-5.
4
Transformer-Based Model with Dynamic Attention Pyramid Head for Semantic Segmentation of VHR Remote Sensing Imagery.基于Transformer且带有动态注意力金字塔头的甚高分辨率遥感影像语义分割模型
Entropy (Basel). 2022 Nov 6;24(11):1619. doi: 10.3390/e24111619.
5
DECTNet: Dual Encoder Network combined convolution and Transformer architecture for medical image segmentation.DECTNet:用于医学图像分割的双编码器网络结合卷积和 Transformer 架构。
PLoS One. 2024 Apr 4;19(4):e0301019. doi: 10.1371/journal.pone.0301019. eCollection 2024.
6
ETU-Net: edge enhancement-guided U-Net with transformer for skin lesion segmentation.ETU-Net:基于边缘增强引导的 U-Net 与 Transformer 的皮肤病变分割。
Phys Med Biol. 2023 Dec 22;69(1). doi: 10.1088/1361-6560/ad13d2.
7
CSformer: Bridging Convolution and Transformer for Compressive Sensing.CSformer:用于压缩感知的桥接卷积和 Transformer。
IEEE Trans Image Process. 2023;32:2827-2842. doi: 10.1109/TIP.2023.3274988. Epub 2023 May 22.
8
MDST: multi-domain sparse-view CT reconstruction based on convolution and swin transformer.MDST:基于卷积和 Swin Transformer 的多域稀疏视图 CT 重建。
Phys Med Biol. 2023 Apr 26;68(9). doi: 10.1088/1361-6560/acc2ab.
9
A novel pansharpening method based on cross stage partial network and transformer.一种基于跨阶段局部网络和变压器的新型全色锐化方法。
Sci Rep. 2024 Jun 2;14(1):12631. doi: 10.1038/s41598-024-63336-w.
10
MESTrans: Multi-scale embedding spatial transformer for medical image segmentation.MESTrans:用于医学图像分割的多尺度嵌入空间变换器
Comput Methods Programs Biomed. 2023 May;233:107493. doi: 10.1016/j.cmpb.2023.107493. Epub 2023 Mar 17.

引用本文的文献

1
Spatiotemporal uncertainty guided non maximum suppression for video event detection.用于视频事件检测的时空不确定性引导非极大值抑制
Sci Rep. 2025 Apr 8;15(1):12019. doi: 10.1038/s41598-025-96188-z.

本文引用的文献

1
A Multi-Scale Recursive Attention Feature Fusion Network for Image Super-Resolution Reconstruction Algorithm.一种用于图像超分辨率重建算法的多尺度递归注意力特征融合网络
Sensors (Basel). 2023 Nov 28;23(23):9458. doi: 10.3390/s23239458.
2
Sparse convolutional neural network for high-resolution skull shape completion and shape super-resolution.用于高分辨率颅骨形状完成和形状超分辨率的稀疏卷积神经网络。
Sci Rep. 2023 Nov 19;13(1):20229. doi: 10.1038/s41598-023-47437-6.
3
Spatial and Channel Aggregation Network for Lightweight Image Super-Resolution.
用于轻量级图像超分辨率的空间与通道聚合网络
Sensors (Basel). 2023 Oct 1;23(19):8213. doi: 10.3390/s23198213.
4
HIPA: Hierarchical Patch Transformer for Single Image Super Resolution.HIPA:用于单图像超分辨率的分层补丁转换器。
IEEE Trans Image Process. 2023;32:3226-3237. doi: 10.1109/TIP.2023.3279977. Epub 2023 Jun 6.
5
GuidedNet: A General CNN Fusion Framework via High-Resolution Guidance for Hyperspectral Image Super-Resolution.GuidedNet:一种基于高分辨率引导的全卷积融合框架的高光谱图像超分辨率方法。
IEEE Trans Cybern. 2023 Jul;53(7):4148-4161. doi: 10.1109/TCYB.2023.3238200. Epub 2023 Jun 15.
6
Decision Fusion Networks for Image Classification.用于图像分类的决策融合网络
IEEE Trans Neural Netw Learn Syst. 2025 Mar;36(3):3890-3903. doi: 10.1109/TNNLS.2022.3196129. Epub 2025 Feb 28.
7
Super-resolution generative adversarial networks with static T2*WI-based subject-specific learning to improve spatial difference sensitivity in fMRI activation.基于静态 T2*WI 的具有个体学习功能的超分辨率生成对抗网络,以提高 fMRI 激活中的空间差异灵敏度。
Sci Rep. 2022 Jun 20;12(1):10319. doi: 10.1038/s41598-022-14421-5.
8
A Non-Local Superpatch-Based Algorithm Exploiting Low Rank Prior for Restoration of Hyperspectral Images.一种基于非局部超补丁且利用低秩先验的高光谱图像恢复算法。
IEEE Trans Image Process. 2021;30:6335-6348. doi: 10.1109/TIP.2021.3093780. Epub 2021 Jul 13.
9
UNet++: A Nested U-Net Architecture for Medical Image Segmentation.U-Net++:一种用于医学图像分割的嵌套U-Net架构。
Deep Learn Med Image Anal Multimodal Learn Clin Decis Support (2018). 2018 Sep;11045:3-11. doi: 10.1007/978-3-030-00889-5_1. Epub 2018 Sep 20.
10
Image super-resolution via sparse representation.基于稀疏表示的图像超分辨率重建。
IEEE Trans Image Process. 2010 Nov;19(11):2861-73. doi: 10.1109/TIP.2010.2050625. Epub 2010 May 18.