• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

TTST:一种用于遥感图像超分辨率的Top-k令牌选择变换器

TTST: A Top-k Token Selective Transformer for Remote Sensing Image Super-Resolution.

作者信息

Xiao Yi, Yuan Qiangqiang, Jiang Kui, He Jiang, Lin Chia-Wen, Zhang Liangpei

出版信息

IEEE Trans Image Process. 2024;33:738-752. doi: 10.1109/TIP.2023.3349004. Epub 2024 Jan 12.

DOI:10.1109/TIP.2023.3349004
PMID:38194374
Abstract

Transformer-based method has demonstrated promising performance in image super-resolution tasks, due to its long-range and global aggregation capability. However, the existing Transformer brings two critical challenges for applying it in large-area earth observation scenes: (1) redundant token representation due to most irrelevant tokens; (2) single-scale representation which ignores scale correlation modeling of similar ground observation targets. To this end, this paper proposes to adaptively eliminate the interference of irreverent tokens for a more compact self-attention calculation. Specifically, we devise a Residual Token Selective Group (RTSG) to grasp the most crucial token by dynamically selecting the top- k keys in terms of score ranking for each query. For better feature aggregation, a Multi-scale Feed-forward Layer (MFL) is developed to generate an enriched representation of multi-scale feature mixtures during feed-forward process. Moreover, we also proposed a Global Context Attention (GCA) to fully explore the most informative components, thus introducing more inductive bias to the RTSG for an accurate reconstruction. In particular, multiple cascaded RTSGs form our final Top- k Token Selective Transformer (TTST) to achieve progressive representation. Extensive experiments on simulated and real-world remote sensing datasets demonstrate our TTST could perform favorably against state-of-the-art CNN-based and Transformer-based methods, both qualitatively and quantitatively. In brief, TTST outperforms the state-of-the-art approach (HAT-L) in terms of PSNR by 0.14 dB on average, but only accounts for 47.26% and 46.97% of its computational cost and parameters. The code and pre-trained TTST will be available at https://github.com/XY-boy/TTST for validation.

摘要

基于Transformer的方法由于其长距离和全局聚合能力,在图像超分辨率任务中展现出了良好的性能。然而,现有的Transformer在将其应用于大面积地球观测场景时带来了两个关键挑战:(1)由于大多数无关令牌导致的冗余令牌表示;(2)忽略了相似地面观测目标的尺度相关性建模的单尺度表示。为此,本文提出自适应消除无关令牌的干扰,以进行更紧凑的自注意力计算。具体而言,我们设计了一个残差令牌选择组(RTSG),通过根据每个查询的分数排名动态选择前k个键来抓住最关键的令牌。为了更好地进行特征聚合,开发了一个多尺度前馈层(MFL),以在前馈过程中生成多尺度特征混合的丰富表示。此外,我们还提出了一种全局上下文注意力(GCA),以充分探索最具信息性的组件,从而为RTSG引入更多归纳偏差以进行准确重建。特别是,多个级联的RTSG形成了我们最终的Top-k令牌选择Transformer(TTST)以实现渐进表示。在模拟和真实世界遥感数据集上的大量实验表明,我们的TTST在定性和定量方面都能优于基于CNN和基于Transformer的现有方法。简而言之,TTST在PSNR方面平均比最先进的方法(HAT-L)高出0.14 dB,但计算成本和参数仅占其47.26%和46.97%。代码和预训练的TTST将在https://github.com/XY-boy/TTST上提供以供验证。

相似文献

1
TTST: A Top-k Token Selective Transformer for Remote Sensing Image Super-Resolution.TTST:一种用于遥感图像超分辨率的Top-k令牌选择变换器
IEEE Trans Image Process. 2024;33:738-752. doi: 10.1109/TIP.2023.3349004. Epub 2024 Jan 12.
2
Efficient Vision Transformer via Token Merger.通过令牌合并实现高效视觉Transformer
IEEE Trans Image Process. 2023;32:4156-4169. doi: 10.1109/TIP.2023.3293763. Epub 2023 Jul 21.
3
Fast-iTPN: Integrally Pre-Trained Transformer Pyramid Network With Token Migration.快速iTPN:带令牌迁移的整体预训练变压器金字塔网络
IEEE Trans Pattern Anal Mach Intell. 2024 Dec;46(12):9766-9779. doi: 10.1109/TPAMI.2024.3429508. Epub 2024 Nov 6.
4
Transforming Image Super-Resolution: A ConvFormer-Based Efficient Approach.变换图像超分辨率:一种基于卷积变换器的高效方法。
IEEE Trans Image Process. 2024;33:6071-6082. doi: 10.1109/TIP.2024.3477350. Epub 2024 Oct 25.
5
Multi-Scale Efficient Graph-Transformer for Whole Slide Image Classification.多尺度高效图Transformer 用于全幻灯片图像分类。
IEEE J Biomed Health Inform. 2023 Dec;27(12):5926-5936. doi: 10.1109/JBHI.2023.3317067. Epub 2023 Dec 5.
6
CSformer: Bridging Convolution and Transformer for Compressive Sensing.CSformer:用于压缩感知的桥接卷积和 Transformer。
IEEE Trans Image Process. 2023;32:2827-2842. doi: 10.1109/TIP.2023.3274988. Epub 2023 May 22.
7
ATTransUNet: An enhanced hybrid transformer architecture for ultrasound and histopathology image segmentation.ATTransUNet:一种用于超声和组织病理学图像分割的增强型混合变压器架构。
Comput Biol Med. 2023 Jan;152:106365. doi: 10.1016/j.compbiomed.2022.106365. Epub 2022 Nov 28.
8
Token Selection is a Simple Booster for Vision Transformers.令牌选择是视觉Transformer的一种简单增强方法。
IEEE Trans Pattern Anal Mach Intell. 2023 Nov;45(11):12738-12746. doi: 10.1109/TPAMI.2022.3208922. Epub 2023 Oct 3.
9
Image Super-Resolution via Efficient Transformer Embedding Frequency Decomposition With Restart.通过带重启的高效变压器嵌入频率分解实现图像超分辨率
IEEE Trans Image Process. 2024;33:4670-4685. doi: 10.1109/TIP.2024.3444317. Epub 2024 Aug 28.
10
Cross-Scope Spatial-Spectral Information Aggregation for Hyperspectral Image Super-Resolution.用于高光谱图像超分辨率的跨范围空间光谱信息聚合
IEEE Trans Image Process. 2024;33:5878-5891. doi: 10.1109/TIP.2024.3468905. Epub 2024 Oct 18.

引用本文的文献

1
Deep learning-based spatio-temporal fusion for high-fidelity ultra-high-speed X-ray radiography.基于深度学习的时空融合用于高保真超高速X射线摄影
J Synchrotron Radiat. 2025 Mar 1;32(Pt 2):432-441. doi: 10.1107/S1600577525000323. Epub 2025 Feb 12.
2
LGIT: local-global interaction transformer for low-light image denoising.LGIT:用于低光照图像去噪的局部-全局交互变换器
Sci Rep. 2024 Sep 18;14(1):21760. doi: 10.1038/s41598-024-72912-z.
3
Research on the prediction of English topic richness in the context of multimedia data.多媒体数据背景下英语话题丰富度预测研究
PeerJ Comput Sci. 2024 Apr 16;10:e1967. doi: 10.7717/peerj-cs.1967. eCollection 2024.