• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

青光眼生成器:用于广义青光眼阶段分类的双域全局Transformer网络

Glaucoformer: Dual-domain Global Transformer Network for Generalized Glaucoma Stage Classification.

作者信息

Das Dipankar, Nayak Deepak Ranjan, Pachori Ram Bilas

出版信息

IEEE J Biomed Health Inform. 2025 May 29;PP. doi: 10.1109/JBHI.2025.3574997.

DOI:10.1109/JBHI.2025.3574997
PMID:40440150
Abstract

Classification of glaucoma stages remains challenging due to substantial inter-stage similarities, the presence of irrelevant features, and subtle lesion size, shape, and color variations in fundus images. For this purpose, few efforts have recently been made using traditional machine learning and deep learning models, specifically convolutional neural networks (CNN). While the conventional CNN models capture local contextual features within fixed receptive fields, they fail to exploit global contextual dependencies. Transformers, on the other hand, are capable of modeling global contextual information. However, they lack the ability to capture local contexts and merely focus on performing attention in the spatial domain, ignoring feature analysis in the frequency domain. To address these issues, we present a novel dual-domain global transformer network, Glaucoformer, to effectively classify glaucoma stages. Specifically, we propose a dual-domain global transformer layer (DGTL) consisting of dual-domain channel attention (DCA) and dual-domain spatial attention (DSA) with Fourier domain feature analyzer (FDFA) as the core component and integrated with a backbone. This helps in exploiting local and global contextual feature dependencies in both spatial and frequency domains, thereby learning prominent and discriminant feature representations. A shared key-query scheme is introduced to learn complementary features while reducing the parameters. In addition, the DGTL leverages the benefits of a deformable convolution to enable the model to handle complex lesion irregularities. We evaluate our method on a benchmark dataset, and the experimental results and extensive comparisons with existing CNN and vision transformer-based approaches indicate its effectiveness for glaucoma stage classification. Also, the results on an unseen dataset demonstrate the generalizability of the model.

摘要

由于青光眼各阶段之间存在显著的相似性、存在无关特征以及眼底图像中病变的大小、形状和颜色变化细微,青光眼阶段的分类仍然具有挑战性。为此,最近很少有人使用传统机器学习和深度学习模型,特别是卷积神经网络(CNN)来进行相关研究。虽然传统的CNN模型能够在固定的感受野内捕捉局部上下文特征,但它们无法利用全局上下文依赖关系。另一方面,Transformer能够对全局上下文信息进行建模。然而,它们缺乏捕捉局部上下文的能力,仅仅专注于在空间域中执行注意力操作,而忽略了频域中的特征分析。为了解决这些问题,我们提出了一种新颖的双域全局Transformer网络Glaucoformer,以有效地对青光眼阶段进行分类。具体来说,我们提出了一种双域全局Transformer层(DGTL),它由双域通道注意力(DCA)和双域空间注意力(DSA)组成,以傅里叶域特征分析器(FDFA)为核心组件,并与主干网络集成。这有助于在空间和频域中利用局部和全局上下文特征依赖关系,从而学习突出的和有区分力的特征表示。引入了一种共享的键值查询方案来学习互补特征,同时减少参数。此外,DGTL利用了可变形卷积的优势,使模型能够处理复杂的病变不规则性。我们在一个基准数据集上评估了我们的方法,实验结果以及与现有基于CNN和视觉Transformer的方法的广泛比较表明了其在青光眼阶段分类中的有效性。此外,在一个未见数据集上的结果证明了该模型的泛化能力。

相似文献

1
Glaucoformer: Dual-domain Global Transformer Network for Generalized Glaucoma Stage Classification.青光眼生成器:用于广义青光眼阶段分类的双域全局Transformer网络
IEEE J Biomed Health Inform. 2025 May 29;PP. doi: 10.1109/JBHI.2025.3574997.
2
A 3D hierarchical cross-modality interaction network using transformers and convolutions for brain glioma segmentation in MR images.一种使用变换和卷积的 3D 层次跨模态交互网络,用于磁共振图像中的脑胶质瘤分割。
Med Phys. 2024 Nov;51(11):8371-8389. doi: 10.1002/mp.17354. Epub 2024 Aug 13.
3
GT-Net: global transformer network for multiclass brain tumor classification using MR images.GT-Net:用于使用磁共振图像进行多类脑肿瘤分类的全局变压器网络。
Biomed Eng Lett. 2024 May 31;14(5):1069-1077. doi: 10.1007/s13534-024-00393-0. eCollection 2024 Sep.
4
GlobalSR: Global context network for single image super-resolution via deformable convolution attention and fast Fourier convolution.GlobalSR:基于可变形卷积注意力和快速傅里叶卷积的单图像超分辨率全局上下文网络。
Neural Netw. 2024 Dec;180:106686. doi: 10.1016/j.neunet.2024.106686. Epub 2024 Aug 31.
5
A spatial-spectral fusion convolutional transformer network with contextual multi-head self-attention for hyperspectral image classification.一种用于高光谱图像分类的具有上下文多头自注意力机制的空间-光谱融合卷积变压器网络。
Neural Netw. 2025 Jul;187:107350. doi: 10.1016/j.neunet.2025.107350. Epub 2025 Mar 14.
6
CVTrack: Combined Convolutional Neural Network and Vision Transformer Fusion Model for Visual Tracking.CVTrack:用于视觉跟踪的卷积神经网络与视觉Transformer融合模型
Sensors (Basel). 2024 Jan 3;24(1):274. doi: 10.3390/s24010274.
7
Dual encoder network with transformer-CNN for multi-organ segmentation.基于 Transformer-CNN 的双编码器网络的多器官分割。
Med Biol Eng Comput. 2023 Mar;61(3):661-671. doi: 10.1007/s11517-022-02723-9. Epub 2022 Dec 29.
8
Enhanced Pneumonia Detection in Chest X-Rays Using Hybrid Convolutional and Vision Transformer Networks.使用混合卷积和视觉Transformer网络增强胸部X光片中的肺炎检测
Curr Med Imaging. 2025;21:e15734056326685. doi: 10.2174/0115734056326685250101113959.
9
A modality-collaborative convolution and transformer hybrid network for unpaired multi-modal medical image segmentation with limited annotations.一种用于具有有限标注的未配对多模态医学图像分割的模态协作卷积与Transformer混合网络。
Med Phys. 2023 Sep;50(9):5460-5478. doi: 10.1002/mp.16338. Epub 2023 Mar 15.
10
TPFR-Net: U-shaped model for lung nodule segmentation based on transformer pooling and dual-attention feature reorganization.TPFR-Net:基于Transformer 池化和双注意力特征重排的肺结节分割 U 型模型。
Med Biol Eng Comput. 2023 Aug;61(8):1929-1946. doi: 10.1007/s11517-023-02852-9. Epub 2023 May 27.