• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于高光谱解混的空间通道多尺度Transformer网络

Spatial-Channel Multiscale Transformer Network for Hyperspectral Unmixing.

作者信息

Sun Haixin, Cao Qiuguang, Meng Fanlei, Xu Jingwen, Cheng Mengdi

机构信息

College of Electronic and Information Engineering, Changchun University, Changchun 130022, China.

出版信息

Sensors (Basel). 2025 Jul 19;25(14):4493. doi: 10.3390/s25144493.

DOI:10.3390/s25144493
PMID:40732622
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12299682/
Abstract

In recent years, deep learning (DL) has been demonstrated remarkable capabilities in hyperspectral unmixing (HU) due to its powerful feature representation ability. Convolutional neural networks (CNNs) are effective in capturing local spatial information, but limited in modeling long-range dependencies. In contrast, transformer architectures extract global contextual features via multi-head self-attention (MHSA) mechanisms. However, most existing transformer-based HU methods focus only on spatial or spectral modeling at a single scale, lacking a unified mechanism to jointly explore spatial and channel-wise dependencies. This limitation is particularly critical for multiscale contextual representation in complex scenes. To address these issues, this article proposes a novel Spatial-Channel Multiscale Transformer Network (SCMT-Net) for HU. Specifically, a compact feature projection (CFP) module is first used to extract shallow discriminative features. Then, a spatial multiscale transformer (SMT) and a channel multiscale transformer (CMT) are sequentially applied to model contextual relations across spatial dimensions and long-range dependencies among spectral channels. In addition, a multiscale multi-head self-attention (MMSA) module is designed to extract rich multiscale global contextual and channel information, enabling a balance between accuracy and efficiency. An efficient feed-forward network (E-FFN) is further introduced to enhance inter-channel information flow and fusion. Experiments conducted on three real hyperspectral datasets (Samson, Jasper and Apex) and one synthetic dataset showed that SCMT-Net consistently outperformed existing approaches in both abundance estimation and endmember extraction, demonstrating superior accuracy and robustness.

摘要

近年来,深度学习(DL)因其强大的特征表示能力,在高光谱解混(HU)中展现出卓越的性能。卷积神经网络(CNNs)在捕捉局部空间信息方面很有效,但在对长距离依赖进行建模时存在局限性。相比之下,Transformer架构通过多头自注意力(MHSA)机制提取全局上下文特征。然而,大多数现有的基于Transformer的HU方法仅专注于单尺度的空间或光谱建模,缺乏一种统一的机制来联合探索空间和通道维度上的依赖关系。这种限制对于复杂场景中的多尺度上下文表示尤为关键。为了解决这些问题,本文提出了一种用于HU的新型空间 - 通道多尺度Transformer网络(SCMT - Net)。具体而言,首先使用一个紧凑特征投影(CFP)模块来提取浅层判别特征。然后,依次应用空间多尺度Transformer(SMT)和通道多尺度Transformer(CMT)来对跨空间维度的上下文关系以及光谱通道之间的长距离依赖进行建模。此外,设计了一个多尺度多头自注意力(MMSA)模块来提取丰富的多尺度全局上下文和通道信息,从而在准确性和效率之间实现平衡。进一步引入了一个高效前馈网络(E - FFN)来增强通道间的信息流和融合。在三个真实高光谱数据集(Samson、Jasper和Apex)以及一个合成数据集上进行的实验表明,SCMT - Net在丰度估计和端元提取方面均持续优于现有方法,展现出卓越的准确性和鲁棒性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/287974f16fc2/sensors-25-04493-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/22ef1999a73d/sensors-25-04493-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/ebb9a686547a/sensors-25-04493-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/ffb1e1ab6d98/sensors-25-04493-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/528896f0b450/sensors-25-04493-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/fc3d11a7bd88/sensors-25-04493-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/9929268840a0/sensors-25-04493-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/0116677fc212/sensors-25-04493-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/30fa83fa892f/sensors-25-04493-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/e7678a3988c0/sensors-25-04493-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/f6ca2fb3e108/sensors-25-04493-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/7a15639b6043/sensors-25-04493-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/03317a6386ac/sensors-25-04493-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/cd746cd62d31/sensors-25-04493-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/287974f16fc2/sensors-25-04493-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/22ef1999a73d/sensors-25-04493-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/ebb9a686547a/sensors-25-04493-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/ffb1e1ab6d98/sensors-25-04493-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/528896f0b450/sensors-25-04493-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/fc3d11a7bd88/sensors-25-04493-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/9929268840a0/sensors-25-04493-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/0116677fc212/sensors-25-04493-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/30fa83fa892f/sensors-25-04493-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/e7678a3988c0/sensors-25-04493-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/f6ca2fb3e108/sensors-25-04493-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/7a15639b6043/sensors-25-04493-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/03317a6386ac/sensors-25-04493-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/cd746cd62d31/sensors-25-04493-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/553e/12299682/287974f16fc2/sensors-25-04493-g014.jpg

相似文献

1
Spatial-Channel Multiscale Transformer Network for Hyperspectral Unmixing.用于高光谱解混的空间通道多尺度Transformer网络
Sensors (Basel). 2025 Jul 19;25(14):4493. doi: 10.3390/s25144493.
2
Multi-level channel-spatial attention and light-weight scale-fusion network (MCSLF-Net): multi-level channel-spatial attention and light-weight scale-fusion transformer for 3D brain tumor segmentation.多级通道空间注意力与轻量级尺度融合网络(MCSLF-Net):用于3D脑肿瘤分割的多级通道空间注意力与轻量级尺度融合变换器
Quant Imaging Med Surg. 2025 Jul 1;15(7):6301-6325. doi: 10.21037/qims-2025-354. Epub 2025 Jun 30.
3
Spectral-spatial wave and frequency interactive transformer for hyperspectral image classification.用于高光谱图像分类的光谱-空间波与频率交互式变压器
Sci Rep. 2025 Jul 26;15(1):27259. doi: 10.1038/s41598-025-12489-3.
4
CLT-MambaSeg: An integrated model of Convolution, Linear Transformer and Multiscale Mamba for medical image segmentation.CLT-MambaSeg:一种用于医学图像分割的卷积、线性变换器和多尺度曼巴的集成模型。
Comput Biol Med. 2025 Sep;196(Pt B):110736. doi: 10.1016/j.compbiomed.2025.110736. Epub 2025 Jul 26.
5
DGCFNet: Dual Global Context Fusion Network for remote sensing image semantic segmentation.DGCFNet:用于遥感图像语义分割的双全局上下文融合网络
PeerJ Comput Sci. 2025 Mar 27;11:e2786. doi: 10.7717/peerj-cs.2786. eCollection 2025.
6
A novel recursive transformer-based U-Net architecture for enhanced multi-scale medical image segmentation.一种基于递归变压器的新型U-Net架构,用于增强多尺度医学图像分割。
Comput Biol Med. 2025 Sep;196(Pt A):110658. doi: 10.1016/j.compbiomed.2025.110658. Epub 2025 Jul 6.
7
Short-Term Memory Impairment短期记忆障碍
8
3D-WDA-PMorph: Efficient 3D MRI/TRUS Prostate Registration using Transformer-CNN Network and Wavelet-3D-Depthwise-Attention.3D-WDA-PMorph:使用Transformer-CNN网络和小波3D深度注意力的高效3D MRI/TRUS前列腺配准
J Imaging Inform Med. 2025 Jul 25. doi: 10.1007/s10278-025-01615-2.
9
Spatio-temporal transformer and graph convolutional networks based traffic flow prediction.基于时空变换器和图卷积网络的交通流预测
Sci Rep. 2025 Jul 7;15(1):24299. doi: 10.1038/s41598-025-10287-5.
10
DASNet a dual branch multi level attention sheep counting network.DASNet是一种双分支多级注意力羊只计数网络。
Sci Rep. 2025 Jul 2;15(1):23228. doi: 10.1038/s41598-025-97929-w.

本文引用的文献

1
CrossFormer++: A Versatile Vision Transformer Hinging on Cross-Scale Attention.CrossFormer++:一种基于跨尺度注意力的通用视觉Transformer
IEEE Trans Pattern Anal Mach Intell. 2024 May;46(5):3123-3136. doi: 10.1109/TPAMI.2023.3341806. Epub 2024 Apr 3.
2
Adversarial Autoencoder Network for Hyperspectral Unmixing.用于高光谱解混的对抗自动编码器网络
IEEE Trans Neural Netw Learn Syst. 2023 Aug;34(8):4555-4569. doi: 10.1109/TNNLS.2021.3114203. Epub 2023 Aug 4.
3
Endmember-Guided Unmixing Network (EGU-Net): A General Deep Learning Framework for Self-Supervised Hyperspectral Unmixing.
端元引导解混网络(EGU-Net):一种用于自监督高光谱解混的通用深度学习框架。
IEEE Trans Neural Netw Learn Syst. 2022 Nov;33(11):6518-6531. doi: 10.1109/TNNLS.2021.3082289. Epub 2022 Oct 27.