Suppr超能文献

通过关注特征对齐实现有效的深度迁移。

Towards effective deep transfer via attentive feature alignment.

机构信息

South China University of Technology, China; PengCheng Laboratory, China.

PengCheng Laboratory, China.

出版信息

Neural Netw. 2021 Jun;138:98-109. doi: 10.1016/j.neunet.2021.01.022. Epub 2021 Feb 10.

Abstract

Training a deep convolutional network from scratch requires a large amount of labeled data, which however may not be available for many practical tasks. To alleviate the data burden, a practical approach is to adapt a pre-trained model learned on the large source domain to the target domain, but the performance can be limited when the source and target domain data distributions have large differences. Some recent works attempt to alleviate this issue by imposing feature alignment over the intermediate feature maps between the source and target networks. However, for a source model, many of the channels/spatial-features for each layer can be irrelevant to the target task. Thus, directly applying feature alignment may not achieve promising performance. In this paper, we propose an Attentive Feature Alignment (AFA) method for effective domain knowledge transfer by identifying and attending on the relevant channels and spatial features between two domains. To this end, we devise two learnable attentive modules at both the channel and spatial levels. We then sequentially perform attentive spatial- and channel-level feature alignments between the source and target networks, in which the target model and attentive module are learned simultaneously. Moreover, we theoretically analyze the generalization performance of our method, which confirms its superiority to existing methods. Extensive experiments on both image classification and face recognition demonstrate the effectiveness of our method. The source code and the pre-trained models are available at https://github.com/xiezheng-cs/AFAhttps://github.com/xiezheng-cs/AFA.

摘要

从头开始训练深度卷积网络需要大量的标记数据,但对于许多实际任务来说,这些数据可能不可用。为了减轻数据负担,一种实用的方法是将在大型源域上学习到的预训练模型适配到目标域,但当源域和目标域的数据分布存在较大差异时,性能可能会受到限制。最近的一些工作试图通过在源和目标网络之间的中间特征图上施加特征对齐来缓解这个问题。然而,对于源模型来说,每个层的许多通道/空间特征可能与目标任务无关。因此,直接应用特征对齐可能无法达到理想的性能。在本文中,我们提出了一种注意力特征对齐(AFA)方法,通过在两个域之间识别和关注相关的通道和空间特征,来实现有效的域知识迁移。为此,我们在通道和空间两个层次上设计了两个可学习的注意力模块。然后,我们在源和目标网络之间依次进行注意力空间和通道级特征对齐,其中目标模型和注意力模块是同时学习的。此外,我们从理论上分析了我们方法的泛化性能,这证实了它优于现有方法。我们在图像分类和人脸识别两个方面都进行了广泛的实验,证明了我们方法的有效性。源代码和预训练模型可在 https://github.com/xiezheng-cs/AFA 上获得。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验