• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

注意力和增强在域适应与泛化中假设转移的重要性

On the Importance of Attention and Augmentations for Hypothesis Transfer in Domain Adaptation and Generalization.

作者信息

Sahay Rajat, Thomas Georgi, Jahan Chowdhury Sadman, Manjrekar Mihir, Popp Dan, Savakis Andreas

机构信息

Rochester Institute of Technology, Rochester, NY 14623, USA.

出版信息

Sensors (Basel). 2023 Oct 12;23(20):8409. doi: 10.3390/s23208409.

DOI:10.3390/s23208409
PMID:37896503
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10611075/
Abstract

Unsupervised domain adaptation (UDA) aims to mitigate the performance drop due to the distribution shift between the training and testing datasets. UDA methods have achieved performance gains for models trained on a source domain with labeled data to a target domain with only unlabeled data. The standard feature extraction method in domain adaptation has been convolutional neural networks (CNNs). Recently, attention-based transformer models have emerged as effective alternatives for computer vision tasks. In this paper, we benchmark three attention-based architectures, specifically vision transformer (ViT), shifted window transformer (SWIN), and dual attention vision transformer (DAViT), against convolutional architectures ResNet, HRNet and attention-based ConvNext, to assess the performance of different backbones for domain generalization and adaptation. We incorporate these backbone architectures as feature extractors in the source hypothesis transfer (SHOT) framework for UDA. SHOT leverages the knowledge learned in the source domain to align the image features of unlabeled target data in the absence of source domain data, using self-supervised deep feature clustering and self-training. We analyze the generalization and adaptation performance of these models on standard UDA datasets and aerial UDA datasets. In addition, we modernize the training procedure commonly seen in UDA tasks by adding image augmentation techniques to help models generate richer features. Our results show that ConvNext and SWIN offer the best performance, indicating that the attention mechanism is very beneficial for domain generalization and adaptation with both transformer and convolutional architectures. Our ablation study shows that our modernized training recipe, within the SHOT framework, significantly boosts performance on aerial datasets.

摘要

无监督域适应(UDA)旨在减轻由于训练数据集和测试数据集之间的分布差异而导致的性能下降。UDA方法已使在有标签数据的源域上训练的模型到仅有无标签数据的目标域上训练的模型的性能得到提升。域适应中的标准特征提取方法一直是卷积神经网络(CNN)。最近,基于注意力的Transformer模型已成为计算机视觉任务的有效替代方案。在本文中,我们将三种基于注意力的架构,即视觉Transformer(ViT)、移位窗口Transformer(SWIN)和双注意力视觉Transformer(DAViT),与卷积架构ResNet、HRNet以及基于注意力的ConvNext进行基准测试,以评估不同骨干网络在域泛化和适应方面的性能。我们将这些骨干架构作为特征提取器纳入用于UDA的源假设转移(SHOT)框架中。SHOT利用在源域中学习到的知识,在没有源域数据的情况下,通过自监督深度特征聚类和自训练来对齐无标签目标数据的图像特征。我们分析了这些模型在标准UDA数据集和航空UDA数据集上的泛化和适应性能。此外,我们通过添加图像增强技术来改进UDA任务中常见的训练过程,以帮助模型生成更丰富的特征。我们的结果表明,ConvNext和SWIN表现最佳,这表明注意力机制对于Transformer和卷积架构的域泛化和适应都非常有益。我们的消融研究表明,在SHOT框架内,我们改进后的训练方法显著提高了航空数据集上的性能。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0b7/10611075/a953eeef08c4/sensors-23-08409-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0b7/10611075/653d9e96ad6c/sensors-23-08409-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0b7/10611075/0695df0bf7ca/sensors-23-08409-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0b7/10611075/449f7ebe971e/sensors-23-08409-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0b7/10611075/85900ab7bb3e/sensors-23-08409-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0b7/10611075/803e19896e00/sensors-23-08409-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0b7/10611075/1491fdb347f9/sensors-23-08409-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0b7/10611075/9fa79a951e0d/sensors-23-08409-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0b7/10611075/a953eeef08c4/sensors-23-08409-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0b7/10611075/653d9e96ad6c/sensors-23-08409-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0b7/10611075/0695df0bf7ca/sensors-23-08409-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0b7/10611075/449f7ebe971e/sensors-23-08409-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0b7/10611075/85900ab7bb3e/sensors-23-08409-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0b7/10611075/803e19896e00/sensors-23-08409-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0b7/10611075/1491fdb347f9/sensors-23-08409-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0b7/10611075/9fa79a951e0d/sensors-23-08409-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0b7/10611075/a953eeef08c4/sensors-23-08409-g008.jpg

相似文献

1
On the Importance of Attention and Augmentations for Hypothesis Transfer in Domain Adaptation and Generalization.注意力和增强在域适应与泛化中假设转移的重要性
Sensors (Basel). 2023 Oct 12;23(20):8409. doi: 10.3390/s23208409.
2
TransConv: Transformer Meets Contextual Convolution for Unsupervised Domain Adaptation.变换器卷积:用于无监督域适应的变换器与上下文卷积相结合
Entropy (Basel). 2024 May 29;26(6):469. doi: 10.3390/e26060469.
3
TransVQA: Transferable Vector Quantization Alignment for Unsupervised Domain Adaptation.TransVQA:用于无监督域适应的可转移向量量化对齐
IEEE Trans Image Process. 2024;33:856-866. doi: 10.1109/TIP.2024.3352392. Epub 2024 Jan 19.
4
Do it the transformer way: A comprehensive review of brain and vision transformers for autism spectrum disorder diagnosis and classification.采用变压器方法:自闭症谱系障碍诊断和分类的脑和视觉变压器的全面综述。
Comput Biol Med. 2023 Dec;167:107667. doi: 10.1016/j.compbiomed.2023.107667. Epub 2023 Nov 3.
5
Unsupervised Domain Adaptation for Medical Image Segmentation Using Transformer With Meta Attention.基于带元注意力的 Transformer 的医学图像分割的无监督领域自适应。
IEEE Trans Med Imaging. 2024 Feb;43(2):820-831. doi: 10.1109/TMI.2023.3322581. Epub 2024 Feb 2.
6
Domain Adaptive and Generalizable Network Architectures and Training Strategies for Semantic Image Segmentation.用于语义图像分割的域自适应和通用网络架构及训练策略
IEEE Trans Pattern Anal Mach Intell. 2024 Jan;46(1):220-235. doi: 10.1109/TPAMI.2023.3320613. Epub 2023 Dec 5.
7
ST-GAN: A Swin Transformer-Based Generative Adversarial Network for Unsupervised Domain Adaptation of Cross-Modality Cardiac Segmentation.ST-GAN:一种基于 Swin Transformer 的生成对抗网络,用于跨模态心脏分割的无监督领域自适应。
IEEE J Biomed Health Inform. 2024 Feb;28(2):893-904. doi: 10.1109/JBHI.2023.3336965. Epub 2024 Feb 5.
8
Source Data-Absent Unsupervised Domain Adaptation Through Hypothesis Transfer and Labeling Transfer.基于假设迁移和标注迁移的源数据缺失无监督领域自适应。
IEEE Trans Pattern Anal Mach Intell. 2022 Nov;44(11):8602-8617. doi: 10.1109/TPAMI.2021.3103390. Epub 2022 Oct 4.
9
HViT: Hybrid vision inspired transformer for the assessment of carotid artery plaque by addressing the cross-modality domain adaptation problem in MRI.HViT:用于评估颈动脉斑块的混合视觉启发式变压器,通过解决MRI中的跨模态域适应问题。
Comput Med Imaging Graph. 2023 Oct;109:102295. doi: 10.1016/j.compmedimag.2023.102295. Epub 2023 Sep 9.
10
Divergence-Agnostic Unsupervised Domain Adaptation by Adversarial Attacks.对抗攻击的无监督领域自适应的分歧不可知论。
IEEE Trans Pattern Anal Mach Intell. 2022 Nov;44(11):8196-8211. doi: 10.1109/TPAMI.2021.3109287. Epub 2022 Oct 4.

本文引用的文献

1
Benchmarking Domain Adaptation Methods on Aerial Datasets.在航空数据集上对域自适应方法进行基准测试。
Sensors (Basel). 2021 Dec 2;21(23):8070. doi: 10.3390/s21238070.
2
Deep High-Resolution Representation Learning for Visual Recognition.用于视觉识别的深度高分辨率表征学习
IEEE Trans Pattern Anal Mach Intell. 2021 Oct;43(10):3349-3364. doi: 10.1109/TPAMI.2020.2983686. Epub 2021 Sep 2.