• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

利用具有动态令牌化的清晰注意力变换器在高光谱图像分类中的潜力。

Leveraging potential of limpid attention transformer with dynamic tokenization for hyperspectral image classification.

作者信息

Yadav Dhirendra Prasad, Kumar Deepak, Jalal Anand Singh, Sharma Bhisham, Liatsis Panos

机构信息

Department of Computer Engineering & Applications, G.L.A. University, Mathurar, Uttar Pradesh, India.

Department of Computer Engineering, NIT Meghalaya, Shillong, Meghalaya, India.

出版信息

PLoS One. 2025 Aug 4;20(8):e0328160. doi: 10.1371/journal.pone.0328160. eCollection 2025.

DOI:10.1371/journal.pone.0328160
PMID:40758709
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12321143/
Abstract

Hyperspectral data consists of continuous narrow spectral bands. Due to this, it has less spatial and high spectral information. Convolutional neural networks (CNNs) emerge as a highly contextual information model for remote sensing applications. Unfortunately, CNNs have constraints in their underlying network architecture in regards to the global correlation of spatial and spectral features, making them less reliable for mining and representing the sequential properties of spectral signatures. In this article, limpid size attention network (LSANet) is proposed, which contains 3D and 2D convolution blocks for enhancement of spatial-spectral features of the hyperspectral image (HSI). In addition, limpid attention block (LAB) is designed to provide a global correlation of the spectral and spatial features through LS attention. Furthermore, the computational costs of LS-attention are less compared to the multi-head self-attention (MHSA) of the classical vision transformer (ViT). In the ViT encoder a conditional position encoding (CPE) module is utilized that dynamically generates tokens from the feature maps to capture a richer contextual representation. The LSANet obtained overall accuracy (OA) of 98.78%, 98.67%, 97.52% and 89.45%, respectively, on the Indian Pines (IP), Pavia University (PU), Salina Valley (SV) and Botswana datasets. Our model's quantitative and qualitative results are considerably better than the classical CNN and transformer-based methods.

摘要

高光谱数据由连续的窄光谱带组成。因此,它具有较少的空间信息和较高的光谱信息。卷积神经网络(CNN)作为一种用于遥感应用的高度上下文信息模型而出现。不幸的是,CNN在其底层网络架构中在空间和光谱特征的全局相关性方面存在限制,这使得它们在挖掘和表示光谱特征的顺序属性时不太可靠。在本文中,提出了清晰尺寸注意力网络(LSANet),它包含3D和2D卷积块,用于增强高光谱图像(HSI)的空间 - 光谱特征。此外,设计了清晰注意力块(LAB),以通过LS注意力提供光谱和空间特征的全局相关性。此外,与经典视觉Transformer(ViT)的多头自注意力(MHSA)相比,LS注意力的计算成本更低。在ViT编码器中,使用了条件位置编码(CPE)模块,该模块从特征图动态生成令牌以捕获更丰富的上下文表示。LSANet在印度松树(IP)、帕维亚大学(PU)、萨利纳山谷(SV)和博茨瓦纳数据集上分别获得了98.78%、98.67%、97.52%和89.45%的总体准确率(OA)。我们模型的定量和定性结果明显优于经典的基于CNN和Transformer的方法。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/d4d029858cc6/pone.0328160.g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/f11e67441078/pone.0328160.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/881dc9ee66cc/pone.0328160.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/7224cf033f67/pone.0328160.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/4c2aa5396bc1/pone.0328160.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/f7c769ed4063/pone.0328160.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/b53c0c228380/pone.0328160.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/46851a90baaa/pone.0328160.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/613ced576dcb/pone.0328160.g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/0714a41dec24/pone.0328160.g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/ae761b30b5f0/pone.0328160.g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/7f938797e087/pone.0328160.g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/b5c6d1640715/pone.0328160.g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/2d859bfa6232/pone.0328160.g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/d4d029858cc6/pone.0328160.g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/f11e67441078/pone.0328160.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/881dc9ee66cc/pone.0328160.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/7224cf033f67/pone.0328160.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/4c2aa5396bc1/pone.0328160.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/f7c769ed4063/pone.0328160.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/b53c0c228380/pone.0328160.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/46851a90baaa/pone.0328160.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/613ced576dcb/pone.0328160.g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/0714a41dec24/pone.0328160.g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/ae761b30b5f0/pone.0328160.g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/7f938797e087/pone.0328160.g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/b5c6d1640715/pone.0328160.g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/2d859bfa6232/pone.0328160.g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/422a/12321143/d4d029858cc6/pone.0328160.g014.jpg

相似文献

1
Leveraging potential of limpid attention transformer with dynamic tokenization for hyperspectral image classification.利用具有动态令牌化的清晰注意力变换器在高光谱图像分类中的潜力。
PLoS One. 2025 Aug 4;20(8):e0328160. doi: 10.1371/journal.pone.0328160. eCollection 2025.
2
TLTNet: A novel transscale cascade layered transformer network for enhanced retinal blood vessel segmentation.TLTNet:一种新颖的跨尺度级联分层Transformer 网络,用于增强视网膜血管分割。
Comput Biol Med. 2024 Aug;178:108773. doi: 10.1016/j.compbiomed.2024.108773. Epub 2024 Jun 25.
3
LGF-Net: A multi-scale feature fusion network for thyroid nodule ultrasound image classification.LGF-Net:一种用于甲状腺结节超声图像分类的多尺度特征融合网络。
J Appl Clin Med Phys. 2025 Aug;26(8):e70149. doi: 10.1002/acm2.70149.
4
A spatial-spectral fusion convolutional transformer network with contextual multi-head self-attention for hyperspectral image classification.一种用于高光谱图像分类的具有上下文多头自注意力机制的空间-光谱融合卷积变压器网络。
Neural Netw. 2025 Jul;187:107350. doi: 10.1016/j.neunet.2025.107350. Epub 2025 Mar 14.
5
Short-Term Memory Impairment短期记忆障碍
6
WSDC-ViT: a novel transformer network for pneumonia image classification based on windows scalable attention and dynamic rectified linear unit convolutional modules.WSDC-ViT:一种基于窗口可扩展注意力和动态整流线性单元卷积模块的新型肺炎图像分类变压器网络。
Sci Rep. 2025 Jul 30;15(1):27868. doi: 10.1038/s41598-025-12117-0.
7
A novel deep learning framework for retinal disease detection leveraging contextual and local features cues from retinal images.一种用于视网膜疾病检测的新型深度学习框架,利用来自视网膜图像的上下文和局部特征线索。
Med Biol Eng Comput. 2025 Feb 7. doi: 10.1007/s11517-025-03314-0.
8
Spectral-spatial wave and frequency interactive transformer for hyperspectral image classification.用于高光谱图像分类的光谱-空间波与频率交互式变压器
Sci Rep. 2025 Jul 26;15(1):27259. doi: 10.1038/s41598-025-12489-3.
9
Application of an electronic tongue and hyperspectral imaging with a CNN-transformer fusion model for rapid detection of botanical origins of honey.基于卷积神经网络-Transformer融合模型的电子舌和高光谱成像技术在蜂蜜植物源快速检测中的应用
Anal Methods. 2025 Jul 31;17(30):6231-6244. doi: 10.1039/d4ay02222j.
10
DC-MSSFF Net: Dule-channel multi-scale spatial-spectral feature fusion network for cholangiocarcinoma pathology high-resolution hyperspectral image segmentation.DC-MSSFF网络:用于胆管癌病理高分辨率高光谱图像分割的双通道多尺度空间-光谱特征融合网络
Comput Methods Programs Biomed. 2025 Sep;269:108905. doi: 10.1016/j.cmpb.2025.108905. Epub 2025 Jun 21.

本文引用的文献

1
A Deep Recurrent Learning-Based Region-Focused Feature Detection for Enhanced Target Detection in Multi-Object Media.一种基于深度循环学习的区域聚焦特征检测,用于增强多目标媒体中的目标检测
Sensors (Basel). 2023 Aug 31;23(17):7556. doi: 10.3390/s23177556.
2
Single-Source Domain Expansion Network for Cross-Scene Hyperspectral Image Classification.用于跨场景高光谱图像分类的单源域扩展网络
IEEE Trans Image Process. 2023;32:1498-1512. doi: 10.1109/TIP.2023.3243853. Epub 2023 Mar 2.
3
Achieving Better Category Separability for Hyperspectral Image Classification: A Spatial-Spectral Approach.
实现高光谱图像分类的更好类别可分性:一种空间 - 光谱方法。
IEEE Trans Neural Netw Learn Syst. 2024 Jul;35(7):9621-9635. doi: 10.1109/TNNLS.2023.3235711. Epub 2024 Jul 8.
4
Rotation-Invariant Attention Network for Hyperspectral Image Classification.用于高光谱图像分类的旋转不变注意力网络。
IEEE Trans Image Process. 2022;31:4251-4265. doi: 10.1109/TIP.2022.3177322. Epub 2022 Jun 29.
5
Central Attention Network for Hyperspectral Imagery Classification.用于高光谱图像分类的中央注意力网络。
IEEE Trans Neural Netw Learn Syst. 2023 Nov;34(11):8989-9003. doi: 10.1109/TNNLS.2022.3155114. Epub 2023 Oct 27.
6
A Supervised Segmentation Network for Hyperspectral Image Classification.一种用于高光谱图像分类的监督分割网络。
IEEE Trans Image Process. 2021;30:2810-2825. doi: 10.1109/TIP.2021.3055613. Epub 2021 Feb 12.